Improve PHP Web App Performance Using Memcached

March 17, 2022
Written by
Reviewed by

Improve PHP Web App Performance Using Memcached

In the first part of this series, I stepped through how to create a Markdown-powered blog in PHP using the Slim Framework. In this, the second part in the series, you're going to learn how to use Memcached with PHP to improve the application's performance.

Let's begin!

Prerequisites

You need the following to follow this tutorial:

Why use caching?

While the initial version works perfectly well, its performance would peak reasonably quickly, because — on every request — the blog data is aggregated from a collection of Markdown files with YAML frontmatter in the application's filesystem, parsing out the article data before the blog data can be rendered.

To reduce complexity, the blog articles contain only a minimum amount of information, that being: the publish date, slug, synopsis, title, main image, categories, and the article's content.

This parsed data is then used to hydrate a series of entities (MarkdownBlog\Entity\BlogItem) that model a blog article. Following entity hydration, any unpublished articles (based on their publish date) are filtered out, with the remaining articles sorted in reverse publish date order.

While reading files from the application's filesystem isn't the most performant approach, it's also not, necessarily, a performance killer. The deployment server may be fitted out with properly optimised and tuned hardware, such as NVMe technology.

However, regardless of the hardware the host uses, and how well it's optimised, executing the full Read > Parse > Hydrate > Render pipeline on each request, is inefficient and wasteful.

It's far more efficient to execute the full pipeline only once, and store the aggregated data in a cache. Then, on every subsequent request, until the cache expires, read the aggregated data and render it. Doing so will result in notably lower page load times, enabling the application to handle far more users, and reduce hosting costs and its impact on the environment.

This is not going to be the most optimised approach, but it will improve application performance by at least 47%.

The refactoring will be broken into three parts:

  1. Refactor the article data aggregation process to allow for easier caching integration.
  2. Add a Memcached service to the application's DI (Dependency Injection) container.
  3. Integrate the Memcached service into the article data aggregation process.

Clone the source code

To get started, clone the application's source code to your development machine, change into the cloned directory, and checkout the 1.0.0 tag by running the following commands.

git clone https://github.com/settermjd/slimphp-markdown-blog.git slim-framework-markdown-blog
cd slim-framework-markdown-blog
git checkout tags/1.0.0 -b 1.0.0

The 1.0.0 tag contains the version of the code from the first part of this series. The application, as it will look by the end of this tutorial, is available in the 2.0.0 tag.

Install the required dependencies

Now, you need to add two additional dependencies, so the application can cache the aggregated blog article data in Memcached; these are:

To install them, run the commands below, in the project's root directory.

composer require \
    laminas/laminas-cache \
    laminas/laminas-cache-storage-adapter-memcached

Integrate caching

If you look at public/index.php, you'll see that the blog data is initially aggregated in a DI service named ContentAggregatorInterface::class.

$container->set(
    ContentAggregatorInterface::class,
    fn() => (new ContentAggregatorFactory())->__invoke([
        'path' => __DIR__ . '/../data/posts',
        'parser' => new Parser(),
    ])
);

Then, it's sorted and filtered in the default route's handler definition.


$app->map(['GET'], '/', function (Request $request, Response $response, array $args) {
    $view = $this->get('view');
    /** @var ContentAggregatorInterface $contentAggregator */
    $contentAggregator = $this->get(ContentAggregatorInterface::class);
    $sorter = new \MarkdownBlog\Sorter\SortByReverseDateOrder();
    $items = $contentAggregator->getItems();
    usort($items, $sorter);
    $iterator = new \MarkdownBlog\Iterator\PublishedItemFilterIterator(
        new ArrayIterator($items)
    );
    return $view->render(
        $response,
        'index.html.twig',
        ['items' => $iterator]
    );
});

This approach works, however, it makes implementing caching problematic because the sorting and filtering processes are run on every request, removing most of the caching benefits.

To refactor the application so that caching can be efficiently implemented will require changes to a number of aspects, but not all that many.

Refactor BlogItem to be serializable

The first thing to do is to refactor MarkdownBlog\Entity\BlogItem so that it is serializable. The reason why is because its publishDate member variable is a DateTime object. That will break serialisation, as many built-in PHP classes, including DateTime, are not serializable.

So, the first things to do are to:

  1. Refactor publishDate to be a string
  2. Refactor the getPublishDate method to return a DateTime object.

To do that, in src/Blog/Entity/BlogItem.php, change publishDate's type to be a string, as in the example below.

private string $publishDate;

Then, refactor the populate method, removing the special initialisation case for publishDate, by replacing it with the code below.


public function populate(array $options = [])
{
    $properties = get_class_vars(__CLASS__);
    foreach ($options as $key => $value) {
        if (array_key_exists($key, $properties) && !empty($value)) {
            $this->$key = $value;
        }
    }
}

Finally, update the getPublishDate method to match the code below, so that it returns a DateTime object initialised with publishDate.

public function getPublishDate(): DateTime
{
    return new DateTime($this->publishDate);
}

Refactor ContentAggregatorFilesystem

The next thing to do is to update src/Blog/ContentAggregator/ContentAggregatorFilesystem.php, to do two things:

  1. Move the item sorting to the getItems() method
  2. Create a new method, named getPublishedItems(), that only retrieves published items.

First, update getItems() to match the version below.

public function getItems(): array
{
    $sorter = new SortByReverseDateOrder();
    usort($this->items, $sorter);

    return $this->items;
}

Then, add the getPublishedItems() method below, after getItems().

public function getPublishedItems(): \Traversable
{
    return new PublishedItemFilterIterator(
        new ArrayIterator($this->items)
    );
}

Then, add the following use statements to the top of the file.

use ArrayIterator;
use MarkdownBlog\Iterator\PublishedItemFilterIterator;
use MarkdownBlog\Sorter\SortByReverseDateOrder;

Refactor the DI container

Now, it's time to refactor the DI container. To do that, in public/index.php, replace the ContentAggregatorInterface::class service with the following definition.

$container->set('articles', function(ContainerInterface $c) {
    /** @var CacheInterface $cache */
    $cache = $c->get('cache');
    if ($cache->has('articles')) {
        return $cache->get('articles');
    }

    $items = (new ContentAggregatorFactory())->__invoke([
        'path' => __DIR__ . '/../data/posts',
            'parser' => new Parser(),
        ]
    );
    $cache->set('articles', $items);

    return $items;
});

Then, add the following use statements to the top of the file.

use Psr\Container\ContainerInterface;
use Psr\SimpleCache\CacheInterface;

The revised code starts off by retrieving the application's cache from a, yet-to-be-defined, service named cache. Then, it checks if the cache has an item named articles, which will contain the aggregated articles.

If so, it returns the cached item. If not, the full content aggregation pipeline is executed, the aggregated BlogItem entities are stored in the application's cache, and then they're returned.

Add a Memcached service to the application's DI container

Next, you need to add the cache service (cache) which the refactored content aggregation service (articles) depends upon. To do that, in public/index.php, add the code below after the articles service.

$container->set('cache', function(ContainerInterface $c): CacheInterface {
    $resourceManager = (new MemcachedResourceManager())
        ->addServer('default', ['host' => '<<MEMCACHED_SERVER_HOSTNAME>>']);

    return new SimpleCacheDecorator(
        new Memcached(
            new MemcachedOptions(
                ['resource_manager' => $resourceManager]
            )
        )
    );
});

Then, replace <<MEMCACHED_SERVER_HOSTNAME>> with the hostname, or IP address, of your Memcached server.

After adding the new DI service, add the following use statements to the top of the file.

use Laminas\Cache\Psr\SimpleCache\SimpleCacheDecorator;
use Laminas\Cache\Storage\Adapter\Memcached;
use Laminas\Cache\Storage\Adapter\MemcachedOptions;
use Laminas\Cache\Storage\Adapter\MemcachedResourceManager;

The code specifies the Memcached server to connect to, in $resourceManager. Then, it initialises a Memcached object, which handles interaction with the Memcached server, passing to it $resourceManager.

Following that, it uses that object to initialise and return a SimpleCacheDecorator object. This object isn't, strictly, necessary. However, doing so provides a rather simplistic interface to the Memcached object, one you might be used to if you've implemented caching previously; i.e., methods such as get, set, has, and delete.

Refactor the default route's handler

The final thing to do, in public/index.php, is to update both route definitions. First, replace the default route with the following code.

$app->map(['GET'], '/', function (Request $request, Response $response, array $args) {
    return $this->get('view')->render(
        $response,
        'index.html.twig',
        ['items' => $this->get('articles')->getPublishedItems()]
    );
});

The route retrieves the articles DI service, calls getPublishedItems() to retrieve the published items, and sets the retrieved items as the items template variable. This makes the route definition a lot more succinct, and doesn't mix several concerns together.

Next, replace the second route definition, with the following code.

$app->map(['GET'], '/item/{slug}', function (Request $request, Response $response, array $args) {
    $view = $this->get('view');
    return $view->render(
        $response,
        'view.html.twig',
        ['item' => $this->get('articles')->findItemBySlug($args['slug'])]
    );
});

This change was required so that it, also, references the articles DI service, instead of the deprecated ContentAggregatorInterface::class service.

Test the changes

With all the changes made, start the application by running the following command in the project's root directory.

php -S 127.0.0.1:8080 -t public

If you prefer to use Docker Compose, which I do, download the docker directory and docker-compose.yml file from the 2.0.0 tag of the project's GitHub repository. Then, start the application by running the following command.

docker-compose up -d

Then, in your browser of choice — before loading the application — open a new tab or window. Then, open the Developer Tools and open the Network tab.

Now, open http://localhost:8080. After the page finishes loading, select the HTML request and, in the right-hand side pane, change to the Timings tab.

Downloaded time in Firefox without caching.

Take note of the "Downloaded" time. With this being the first request, no cached content was available, so the full content aggregation process would have been completed, and the result stored in Memcached.

Now, reload the page. The download time should be notably less than the original request, content was retrieved from the cache, not aggregated.

Downloaded time in Firefox with caching.

By using caching, the second page load is about 75% faster than the first load, where the cache was not warmed up.

That's how to improve PHP web app performance using Memcached

Sometimes, as in this application, despite the performance benefits that it brings, caching isn't immediately possible. However, allowing for your particular use-case, with a modest amount of refactoring, you can take advantage of it, and start handling significantly more visitors.

Keen to improve the performance of the application even further? Here are five excellent ways.

Matthew Setter is the PHP Editor in the Twilio Voices team and (naturally) a PHP developer. He’s also the author of Deploy With Docker Compose. When he’s not writing PHP code, he’s editing great PHP articles here at Twilio. You can find him at msetter@twilio.com, Twitter, and GitHub.