{"name":"py311-scrapy","portdir":"python/py-scrapy","version":"2.13.3","license":"BSD","platforms":"{darwin any}","epoch":1,"replaced_by":null,"homepage":"https://scrapy.org/","description":"A high-level Web Crawling and Web Scraping framework","long_description":"Scrapy is a high level scraping and web crawling framework for writing spiders to crawl and parse web pages for all kinds of purposes, from information retrieval to monitoring or testing web sites.","active":true,"categories":["python"],"maintainers":[{"name":"khindenburg","github":"kurthindenburg","ports_count":433}],"variants":[],"dependencies":[{"type":"build","ports":["clang-18","py311-build","py311-installer","py311-setuptools","py311-wheel"]},{"type":"lib","ports":["py311-openssl","py311-service_identity","py311-tldextract","py311-defusedxml","py311-cssselect","py311-dispatcher","py311-itemadapter","py311-itemloaders","py311-parsel","py311-protego","py311-queuelib","py311-twisted","py311-w3lib","python311","py311-typing_extensions","py311-zopeinterface","py311-lxml","py311-cryptography"]},{"type":"run","ports":["scrapy_select"]},{"type":"test","ports":["py311-pytest","py311-Pillow","py311-pexpect","py311-tldextract","py311-pytest-asyncio","py311-botocore","py311-jmespath","py311-sybil","py311-testfixtures"]}],"depends_on":[{"type":"test","ports":["py311-itemloaders","py311-itemadapter"]}]}