This package is a slightly overengineered Diffbot API wrapper. It uses PSR-7 and PHP-HTTP friendly client implementations to make API calls. To learn more about Diffbot see here and their homepage. Right now it only supports Analyze, Product, Image, Discussion, Crawl, Search, and Article APIs, but can also accommodate Custom APIs. Video and Bulk API support coming soon.
Full documentation available here.
Minimum PHP 5.6 is required. PHP 7.0 is recommended.
This package uses some non-stable packages, so you must set your project's minimum stability to something like beta or dev in composer.json
:
"minimum-stability": "dev",
"prefer-stable": true
If you don't the installation procedure below will fail.
The library depends on an implementation of the client-implementation virtual package. If you don't know what this means, simply requiring the Guzzle6 adapter will do:
composer require php-http/guzzle6-adapter
This adapter satisfies the requirement for client-implementation (see above) and will make it possible to install the client with:
composer require swader/diffbot-php-client
Simplest possible use case:
$diffbot = new Diffbot('my_token');
$url = 'http://www.sitepoint.com/diffbot-crawling-visual-machine-learning/';
$articleApi = $diffbot->createArticleAPI($url);
echo $articleApi->call()->author; // prints out "Bruno Skvorc"
That's it, this is all you need to get started.
Full API reference manual in progress, but the instructions below should do for now - the library was designed with brutal UX simplicity in mind.
To begin, always create a Diffbot instance. A Diffbot instance will spawn API instances. To get your token, sign up at http://diffbot.com
$diffbot = new Diffbot('my_token');
Then, pick an API.
Currently available automatic APIs are:
- product (crawls products and their reviews, if available)
- article (crawls news posts, blogs, etc, with comments if available)
- image (fetches information about images - useful for 500px, Flickr etc). The Image API can return several images - depending on how many are on the page being crawled.
- discussion (fetches discussion / review / comment threads - can be embedded in the Product or Article return data, too, if those contain any comments or discussions)
- analyze (combines all the above in that it automatically determines the right API for the URL and applies it)
Video is coming soon. See below for instructions on Crawlbot, Search and Bulk API.
There is also a Custom API like this one - unless otherwise configured, they return instances of the Wildcard entity)
All APIs can also be tested on http://diffbot.com
The API you picked can be spawned through the main Diffbot instance:
$api = $diffbot->createArticleAPI($url);
All APIs have some optional fields you can pass with parameters. For example, to extract the 'meta' values of the page alongside the normal data, call setMeta
:
$api->setMeta(true);
Some APIs have other flags that don't qualify as fields. For example, the Article API can be told to ignore Discussions (aka to not extract comments). This can speed up the fetching, because by default, it does look for them. The configuration methods all have the same format, though, so to accomplish this, just use setDiscussion
:
$api->setDiscussion(false);
All config methods are chainable:
$api->setMeta(true)->setDiscussion(false);
For an overview of all the config fields and the values each API returns, see here.
All API instances have the call
method which returns a collection of results. The collection is iterable:
$url = 'http://smittenkitchen.com/blog/2012/01/buckwheat-baby-with-salted-caramel-syrup/';
$imageApi = $diffbot->createImageAPI($url);
/** @var Image $imageEntity */
foreach ($imageApi->call() as $imageEntity) {
echo 'Image dimensions: ' . $imageEntity->getHeight() . ' x ' . $imageEntity->getWidth() . '<br>';
}
/* Output:
Image dimensions: 333 x 500
Image dimensions: 333 x 500
Image dimensions: 334 x 500
Image dimensions: 333 x 500
Image dimensions: 333 x 500
Image dimensions: 333 x 500
Image dimensions: 333 x 500
Image dimensions: 333 x 500
Image dimensions: 333 x 500
*/
In cases where only one entity is returned, like Article or Product, iterating works all the same, it just iterates through the one single element. The return data is always a collection!
However, for brevity, you can access properties directly on the collection, too.
$articleApi = $diffbot->createArticleAPI('http://www.sitepoint.com/diffbot-crawling-visual-machine-learning/');
echo $articleApi->call()->author;
// or $articleApi->call()->getAuthor();
In this case, the collection applies the property call to the first element which, coincidentally, is also the only element. If you use this approach on the image collection above, the same thing happens - but the call is only applied to the first image entity in the collection.
If you just want the final generated URL (for example, to paste into Postman Client or to test in the browser and get pure JSON), use buildUrl
:
$url = $articleApi->buildUrl();
You can continue regular API usage afterwards, which makes this very useful for logging, etc.
You can extract the pure, full Guzzle Response object from the returned data and then manipulate it as desired (maybe parsing it as JSON and processing it further on your own):
$articleApi = $diffbot->createArticleAPI('http://www.sitepoint.com/diffbot-crawling-visual-machine-learning/');
$guzzleResponse = $articleApi->call()->getResponse();
Individual entities do not have access to the response - to fetch it, always fetch from their parent collection (the object that the call()
method returns).
The Discussion API returns some data about the discussion and contains another collection of Posts. A Post entity corresponds to a single review / comment / forum post, and is very similar in structure to the Article entity.
You can iterate through the posts as usual:
$url = 'http://community.sitepoint.com/t/php7-resource-recap/174325/';
$discussion = $diffbot->createDiscussionAPI($url)->call();
/** @var Post $post */
foreach($discussion->getPosts() as $post) {
echo 'Author: '.$post->getAuthor().'<br>';
}
/*
Output:
Author: swader
Author: TaylorRen
Author: s_molinari
Author: s_molinari
Author: swader
Author: s_molinari
Author: swader
Author: s_molinari
Author: swader
Author: s_molinari
Author: TomB
Author: s_molinari
Author: TomB
Author: Wolf_22
Author: swader
Author: swader
Author: s_molinari
*/
An Article or Product entity can contain a Discussion entity. Access it via getDiscussion
on an Article or Product entity and use as usual (see above).
Used just like all others. There are only two differences:
- When creating a Custom API call, you need to pass in the API name
- It always returns Wildcard entities which are basically just value objects containing the returned data. They have
__call
and__get
magic methods defined so their properties remain just as accessible as the other Entities', but without autocomplete.
The following is a usage example of my own custom API for author profiles at SitePoint:
$diffbot = new Diffbot('my_token');
$customApi = $diffbot->createCustomAPI('http://sitepoint.com/author/bskvorc', 'authorFolioNew');
$return = $customApi->call();
foreach ($return as $wildcard) {
dump($wildcard->getAuthor()); // Bruno Skvorc
dump($wildcard->author); // Bruno Skvorc
}
Of course, you can easily extend the basic Custom API class and make your own, as well as add your own Entities that perfectly correspond to the returned data. This will all be covered in a tutorial in the near future.
Basic Crawlbot support has been added to the library. To find out more about Crawlbot and what, how and why it does what it does, see here. I also recommend reading the Crawlbot API docs and the Crawlbot support topics just so you can dive right in without being too confused by the code below.
In a nutshell, the Crawlbot crawls a set of seed URLs for links (even if a subdomain is passed to it as seed URL, it still looks through the entire main domain and all other subdomains it can find) and then processes all the pages it can find using the API you define (or opting for Analyze API by default).
A joint list of all your crawl / bulk jobs can be fetched via:
$diffbot = new Diffbot('my_token');
$jobs = $diffbot->crawl()->call();
This returns a collection of all crawl and bulk jobs. Each type is represented by its own class: JobCrawl
and JobBulk
. It's important to note that Jobs only contain the information about the job - not the data. To get the data of a job, use the downloadUrl
method to get the URL to the dataset:
$url = $job->downloadUrl("json");
See inline comments for step by step explanation
// Create new diffbot as usual
$diffbot = new Diffbot('my_token');
// The crawlbot needs to be told which API to use to process crawled pages. This is optional - if omitted, it will be told to use the Analyze API with mode set to auto.
// The "crawl" url is a flag to tell APIs to prepare for consumption with Crawlbot, letting them know they won't be used directly.
$url = 'crawl';
$articleApi = $diffbot->createArticleAPI($url)->setDiscussion(false);
// Make a new crawl job. Optionally, pass in API instance
$crawl = $diffbot->crawl('sitepoint_01', $articleApi);
// Set seeds - seeds are URLs to crawl. By default, passing a subdomain into the crawl will also crawl other subdomains on main domain, including www.
$crawl->setSeeds(['http://sitepoint.com']);
// Call as usual - an EntityIterator collection of results is returned. In the case of a job's creation, only one job entity will always be returned.
$job = $crawl->call();
// See JobCrawl class to find out which getters are available
dump($job->getDownloadUrl("json")); // outputs download URL to JSON dataset of the job's result
To get data about a job (this will be the data it was configured with - its flags - and not the results!), use the exact same approach as if creating a new one, only without the API and seeds:
$diffbot = new Diffbot('my_token');
$crawl = $diffbot->crawl('sitepoint_01');
$job = $crawl->call();
dump($job->getDownloadUrl("json")); // outputs download URL to JSON dataset of the job's result
While there is no way to alter a crawl job's configuration post creation, you can still do some operations on it.
Provided you fetched a $crawl
instance as in the above section on inspecting, you can do the following:
// Force start of a new crawl round manually
$crawl->roundStart();
// Pause or unpause (0) a job
$crawl->pause();
$crawl->pause(0)
// Restart removes all crawled data but keeps the job (and settings)
$crawl->restart();
// Delete a job and all related data
$crawl->delete();
Note that it is not necessary to issue a call()
after these methods.
If you would like to extract the generated API call URL for these instant-call actions, pass in the parameter false
, like so:
$crawl->delete(false);
You can then save the URL for your convenience and call call
when ready to execute (if at all).
$url = $crawl->buildUrl();
$url->call();
The Search API is used to quickly search across data obtained through Bulk or Crawl API.
$diffbot = new Diffbot('my_token');
$search = $diffbot->search('author:"Miles Johnson" AND type:article')->call();
foreach ($search as $article) {
echo $article->getTitle();
}
Use Search APIs setCol
method to target a specific collection only - otherwise, all your token's collections are searched.
Just run PHPUnit in the root folder of the cloned project.
Some calls do require an internet connection (see tests/Factory/EntityTest
).
phpunit
I'll pay $10 for every new set of 5 Entity tests, submissions verified set per set - offer valid until I feel like there's enough use cases covered. (a.k.a. don't submit 1500 of them at once, I can't pay that in one go).
If you would like to contribute by adding Entity tests, I suggest following this procedure:
-
Pick an API you would like to contribute a test for. E.g., Product API.
-
In a scratchpad like
index.php
, build the URL:$diffbot = new Diffbot('my_token'); $url = $diffbot ->createProductAPI('http://someurl.com') ->setMeta(true) ->...(insert other config methods here as desired)... ->buildUrl(); echo $url;
-
Grab the URL and paste it into a REST client like Postman or into your browser. You'll get Diffbot's response back. Keep it open for reference.
-
Download this response into a JSON file. Preferably into
tests/Mocks/Products/[date]/somefilename.json
, like the other tests are. This is easily accomplished by executingcurl "[url] > somefilename.json"
in the Terminal/Command Line. -
Go into the appropriate tests folder. In this case,
tests/Entity
and openProductTest.php
. Notice how the file is added into the batch of files to be tested against. Every provider has it referenced, along with the value the method being tested should produce. Slowly go through every test method and add your file. Use the values in the JSON you got in step 3 to get the values. -
Run
phpunit tests/Entity/ProductTest.php
to test just this file (much faster than entire suite). If OK, send PR :)
If you'd like to create your own Test classes, too, that's fine, no need to extend the ones that are included with the project. Apply the whole process just rather than extending the existing ProductTest
class make a new one.
Other tests don't have specific instructions, contribute as you see fit. Just try to minimize actual remote calls - we're not testing the API itself (a.k.a. Diffbot), we're testing this library. If the library parses values accurately from an inaccurate API response because, for example, Diffbot is currently bugged, that's fine - the library works!
Please see CONTRIBUTING for details and TODO for ideas.
The MIT License (MIT). Please see License File for more information.