Metadata-Version: 1.1
Name: solidscraper
Version: 0.7.4
Summary: This package lets your script scrape web sites. JQuery-Like API.
Home-page: https://github.com/sergioburdisso/solidscraper
Author: Sergio Burdisso
Author-email: sergio.burdisso@gmail.com
License: MIT
Download-URL: https://github.com/sergioburdisso/solidscraper/tarball/v0.7.4
Description-Content-Type: UNKNOWN
Description: Solid Scraper
        =============
        
        Easy to use JQuery-Like API for Web Scraping/Crawling. It also supports
        Cookies and custom User Agents.
        
        +-----+
        | ##  |
        | 1.  |
        | Ins |
        | tal |
        | lat |
        | ion |
        +-----+
        | ### |
        | 1.1 |
        | Usi |
        | ng  |
        | pip |
        +-----+
        | ``p |
        | ip  |
        | ins |
        | tal |
        | l s |
        | oli |
        | dsc |
        | rap |
        | er` |
        | `   |
        +-----+
        
        2. "Hello World" Example
        ------------------------
        
        Getting all url of all links:
        
        .. code:: python
        
            import solidscraper as ss
        
            doc = ss.load("https://www.example.com/the/path")
        
            # print the list of urls from all <a> elements
            print doc.select("a").getAttribute("href")
        
        Getting all url of all links inside <div>s whose class id is 'links':
        
        .. code:: python
        
            import solidscraper as ss
        
            doc = ss.load("https://www.example.com/the/path")
        
            # print the list of urls from all <a> elements inside <div id="links">
            print doc.select("div #links").then("a").getAttribute("href")
        
        Getting the text of all <span> elements inside <p> whose class are
        'info':
        
        .. code:: python
        
            import solidscraper as ss
        
            doc = ss.load("https://www.example.com/the/path")
        
            # print the text of all <span> elements inside <p class="info">
            print doc.select("p .info").then("span").text()
        
Keywords: scrape,crawler,crawl,scraper
Platform: UNKNOWN
Classifier: Topic :: Internet :: WWW/HTTP
Classifier: Programming Language :: Python :: Implementation :: PyPy
Classifier: Programming Language :: Python
Classifier: Programming Language :: Python :: 2
Classifier: Programming Language :: Python :: 3
