Metadata-Version: 2.1
Name: scrapy-GUI
Version: 1.0.0
Summary: A package for offering UI tools for building scrapy queries
Home-page: https://github.com/further-reading/scraping-browser
Author: Roy Healy
Author-email: roy.healy87@gmail.com
License: UNKNOWN
Platform: UNKNOWN
Classifier: Programming Language :: Python :: 3
Classifier: License :: OSI Approved :: MIT License
Classifier: Operating System :: OS Independent
Requires-Python: >=3.6
Description-Content-Type: text/markdown

Requires Python 3.6+

# Scrapy UI
A simple, Qt-Webengine powered web browser with built in functionality for testing scrapy spider code.

Also includes an addon to enable a GUI for use with the scrapy shell.


**Table of Contents**

- [Installation](#installation)
- [Standalone UI](#standalone-ui)
    - [Browser Tab](#browser-tab)
    - [Tools Tab](#tools-tab)
        - [Query Box](#query-box)
        - [Regex Box](#regex-box)
        - [Function Box](#function-box)
        - [Results Box](#results-box)
    - [Source Tab](#source-tab)
    - [Notes Tab](#notes-tab)
- [Integration with Scrapy Shell](#integration-with-scrapy-shell)
    - [Activation](#activation)

# Installation

You can import the package from PyPi using

> pip install scrapy_gui

Then you can import it to a shell using `scrapy_ui`

# Standalone UI
The standlaone UI can be opened by using `from scrapy_ui.open_browser()` from a python shell. This consists of a web browser and a set of tools to analyse its contents.

## Browser Tab
Enter any url into search bar and hit return or press the Go button. When the loading animation finishes it will be ready to parse in the Tools tab.

![Browser tab](https://raw.githubusercontent.com/further-reading/scraping-browser/master/readme_images/browser.png "Browser Example")

## Tools Tab
The tools tab contains various sections for parsing content of the page. The purpose of this tab is to make it easy to test queries and code for use in a scrapy spider.
> **NOTE:** This will use the **initial** html response. If additional requests, javascript, etc alter the page later this will not be taken into account.

It will load the initial html with an additional request using the `requests` package. When running a query it will create a selector object using `Selection` from the parsel package.

![Tools tab](https://raw.githubusercontent.com/further-reading/scraping-browser/master/readme_images/tools.png "Tools Example")

### Query Box
The query box lets you use [parsel](https://github.com/scrapy/parsel) compatible CSS queries to extract data from the page.

It returns results as though `selection.css('YOUR QUERY').getall()` was called.

If there are no results or there is an error in the query a dialogue will pop up informing you of the issue.

### Regex Box
This box lets you add a regular expression pattern to be used in addition to the previous css query. 

It returns results as though `selection.css('YOUR QUERY').re(r'YOUR REGEX')'` was called. This means that if you use groups it will only return the content within parenthesis.

### Function Box
This box lets you define additional python code that can run on the results of your query and regex. The code can be as long and complex as you want, including adding additional functions, classes, imports etc.

The only requirement is you must include a function called `user_fun(results, selector)` that returns a `list`. 

### Results Box

This table will list all the results, passed through the regex and function if defined.

## Source Tab

This tab contains the html source that is used in the Tools tab. You can use the text box to search for specific content. All searches are not case sensitive.

![Source Tab](https://raw.githubusercontent.com/further-reading/scraping-browser/master/readme_images/source.png "Source Example")

## Notes Tab

This is just a plain text box. Content in here is not saved when you exit the app.

# Integration with Scrapy Shell

It is possible to integrate this tool with the scrapy shell. This will allow you to use it on responses that have been passed through your middlewares, access more complex requests and more specific selectors.

## Activation

To use it in your shell import the load_selector method using:

`from scrapy_ui import load_selector`

Then you can write load_selector(YOUR_SELECTOR) to open a window with your selector loaded into it.

> For example `load_selector(response)` will load your response into the UI.

When you run the code a window named `Shell UI` will open that contains the `Tools`, `Source` and `Notes` tabs from the standalone window mentioned above.


