{"name":"py310-scrapy","portdir":"python/py-scrapy","version":"2.13.3","license":"BSD","platforms":"{darwin any}","epoch":1,"replaced_by":null,"homepage":"https://scrapy.org/","description":"A high-level Web Crawling and Web Scraping framework","long_description":"Scrapy is a high level scraping and web crawling framework for writing spiders to crawl and parse web pages for all kinds of purposes, from information retrieval to monitoring or testing web sites.","active":true,"categories":["python"],"maintainers":[{"name":"khindenburg","github":"kurthindenburg","ports_count":433}],"variants":[],"dependencies":[{"type":"build","ports":["clang-18","py310-setuptools","py310-wheel","py310-build","py310-installer"]},{"type":"lib","ports":["py310-lxml","py310-cryptography","py310-openssl","py310-twisted","py310-tldextract","py310-service_identity","py310-dispatcher","py310-itemadapter","py310-itemloaders","py310-parsel","py310-protego","py310-queuelib","py310-w3lib","python310","py310-defusedxml","py310-typing_extensions","py310-zopeinterface","py310-cssselect"]},{"type":"run","ports":["scrapy_select"]},{"type":"test","ports":["py310-pytest","py310-Pillow","py310-botocore","py310-jmespath","py310-testfixtures","py310-pexpect","py310-pytest-asyncio","py310-sybil","py310-tldextract"]}],"depends_on":[{"type":"test","ports":["py310-itemloaders","py310-itemadapter"]}]}