Download the PHP
package ssola/crawly without Composer
On this page you can find all versions of the php package
ssola/crawly. It is possible to download/install
these versions without Composer. Possible dependencies are resolved
automatically.
Vendor ssola Package crawly Short Description Simple web crawler library License
FAQ
After the download, you have to make one include require_once('vendor/autoload.php');. After that you have to import the classes with use statements.
Example:
If you use only one package a project is not needed. But if you use more then one package, without a project it is not possible to import the classes with use statements.
In general, it is recommended to use always a project to download your libraries. In an application normally there is more than one library needed.
Some PHP packages are not free to download and because of that hosted in private repositories.
In this case some credentials are needed to access such packages.
Please use the auth.json textarea to insert credentials, if a package is coming from a private repository.
You can look here for more information.
Some hosting areas are not accessible by a terminal or SSH. Then it is not possible to use Composer.
To use Composer is sometimes complicated. Especially for beginners.
Composer needs much resources. Sometimes they are not available on a simple webspace.
If you are using private repositories you don't need to share your credentials. You can set up everything on our site and then you provide a simple download link to your team member.
Simplify your Composer build process. Use our own command line tool to download the vendor folder as binary. This makes your build process faster and you don't need to expose your credentials for private repositories.
Crawly is a simple web crawler able to extract and follow links depending on the discovers.
Simple Example
Crawler object
You can create a simple crawler with the Crawler Factory, it will generate a Crawly object using Guzzle as Http client.
You can create a personalized crawler specified which Http client, Url queue and Visited link collection to use.
Discovers
Discovers are used to extract from the html a set of links to include to the queue. You can include as many discovers as you want and you can create your own discovers classes too.
At the moment Crawly only includes a Css Selector discover.
Create your own discover
Just create a new class that implements the Discoverable interface. This new class should look like this example:
Limiters
Limiters are used to limit the crawler actions. For instance, we can limit how many links can been crawled or which is the maximum amout of bandwitdth to use.
Composer command for our command line client (download client)This client runs in each environment. You don't need a specific PHP version etc. The first 20 API calls are free.Standard composer command
The package ssola/crawly contains the following files
Loading the files please wait ....
Loading please wait ...
Before you can download the PHP files, the dependencies should be resolved. This can take some minutes. Please be patient.