storefrontbvba / zend-cache-redis
Zend_Cache backend using Redis with full support for tags, but without the need for Magento.
Requires
This package is auto-updated.
Last update: 2024-12-06 23:14:47 UTC
README
This Zend_Cache backend allows you to use a Redis server as a central cache storage. Tags are fully supported without the use of TwoLevels cache so this backend is great for use on a single machine or in a cluster. Works with any Zend Framework project!
FEATURES
- Uses the phpredis PECL extension for best performance (requires master branch or tagged version newer than Aug 19 2011).
- Falls back to standalone PHP if phpredis isn't available using the Credis library.
- Tagging is fully supported, implemented using the Redis "set" and "hash" datatypes for efficient tag management.
- Key expiration is handled automatically by Redis.
- Supports unix socket connection for even better performance on a single machine.
- Supports configurable compression for memory savings. Can choose between gzip, lzf and snappy and can change configuration without flushing cache.
- Uses transactions to prevent race conditions between saves, cleans or removes causing unexpected results.
- Supports a configurable "auto expiry lifetime" which, if set, will be used as the TTL when the key otherwise wouldn't expire.
- Unit tested!
High Availability and Load Balancing Support
There are two supported methods of achieving High Availability and Load Balancing with Cm_Cache_Backend_Redis.
Redis Sentinel
You may achieve high availability and load balancing using Redis Sentinel. To enable use of Redis Sentinel the server
specified should be a comma-separated list of Sentinel servers and the sentinel_master
option should be specified
to indicate the name of the sentinel master set (e.g. 'mymaster'). If using sentinel_master
you may also specify
load_from_slaves
in which case a random slave will be chosen for performing reads in order to load balance across multiple Redis instances.
Using the value '1' indicates to only load from slaves and '2' to include the master in the random read slave selection.
Example configuration:
<!-- This is a child node of config/global -->
<cache>
<backend>Cm_Cache_Backend_Redis</backend>
<backend_options>
<server>tcp://10.0.0.1:26379,tcp://10.0.0.2:26379,tcp://10.0.0.3:26379</server>
<timeout>0.5</timeout>
<sentinel_master>mymaster</sentinel_master>
<sentinel_master_verify>1</sentinel_master_verify>
<load_from_slaves>1</load_from_slaves>
</backend_options>
</cache>
Load Balancer or Service Discovery
It is also possible to achieve high availability by using other methods where you can specify separate connection addresses for the
master and slave(s). The load_from_slave
option has been added for this purpose and this option does not
connect to a Sentinel server as the example above, although you probably would benefit from still having a Sentinel setup purely for
the easier replication and failover.
Examples would be to use a TCP load balancer (e.g. HAProxy) with separate ports for master and slaves, or a DNS-based system that uses service discovery health checks to expose master and slaves via different DNS names.
Example configuration:
<!-- This is a child node of config/global -->
<cache>
<backend>Cm_Cache_Backend_Redis</backend>
<backend_options>
<server>tcp://redis-master:6379</server>
<load_from_slave>tcp://redis-slaves:6379</load_from_slave>
<master_write_only>0</master_write_only> <!-- Use 1 to prevent reads from master -->
<timeout>0.5</timeout>
</backend_options>
</cache>
Static Configuration
You may also statically specify the master and slave servers by passing either an array to load_from_slave
or a string
with multiple addresses separated by a comma.
<!-- This is a child node of config/global -->
<cache>
<backend>Cm_Cache_Backend_Redis</backend>
<backend_options>
<server>tcp://redis-master:6379</server>
<load_from_slave>tcp://redis-slave1:6379,tcp://redis-slave2:6379</load_from_slave>
<master_write_only>0</master_write_only> <!-- Use 1 to prevent reads from master -->
<timeout>0.5</timeout>
</backend_options>
</cache>
ElastiCache
The following example configuration lets you use ElastiCache Redis (cluster mode disabled) where the writes are sent to the Primary node and reads are sent to the replicas. This lets you distribute the read traffic between the different nodes.
The instructions to find the primary and read replica endpoints are here.
<!-- This is a child node of config/global/cache -->
<backend_options>
<server>primary-endpoint.0001.euw1.cache.amazonaws.com</server>
<port>6379</port>
<database>0</database> <!-- Make sure database is 0 -->
<master_write_only>1</master_write_only>
<load_from_slave>
<node-001>
<server>replica-endpoint-1.jwbaun.0001.euw1.cache.amazonaws.com</server>
<port>6379</port>
</node-001>
<node-002>
<server>replica-endpoint-2.jwbaun.0001.euw1.cache.amazonaws.com</server>
<port>6379</port>
</node-002>
</load_from_slave>
</backend_options>
DEPRECATION NOTICE
Previously the ElastiCache config instructions suggested setting up a <cluster>
node but this functionality was flawed
and is no longer supported. The config is still parsed and loaded for backwards-compatibility but chooses a random slave
to read from rather than using md5 hash of the keys.
RELATED / TUNING
- The recommended "maxmemory-policy" is "volatile-lru". All tag metadata is non-volatile so it is recommended to use key expirations unless non-volatile keys are absolutely necessary so that tag data cannot get evicted. So, be sure that the "maxmemory" is high enough to accommodate all of the tag data and non-volatile data with enough room left for the volatile key data as well.
- Automatic cleaning is optional and not recommended since it is slow and uses lots of memory.
- Occasional (e.g. once a day) garbage collection is recommended if the entire cache is infrequently cleared and automatic cleaning is not enabled. The best solution is to run a cron job which does the garbage collection. (See "Example Garbage Collection Script" below.)
- Compression will have additional CPU overhead but may be worth it for memory savings and reduced traffic.
For high-latency networks it may even improve performance.
- gzip — Slowest but highest compression. Most likely you will not want to use above level 1 compression.
- lzf — Fastest compress, fast decompress. Install:
sudo pecl install lzf
- snappy — Fastest decompress, fast compress. Download and install: snappy and php-snappy
- Monitor your redis cache statistics with my modified munin plugin.
- Enable persistent connections. Make sure that if you have multiple configurations connecting the persistent string is unique for each configuration so that "select" commands don't cause conflicts.
- Use the
stats.php
script to inspect your cache to find oversized or wasteful cache tags.
Release Notes
- March 2017: Added support for Redis Sentinel and loading from slaves. Thanks @Xon for the help!
- Sometime in 2013: Ceased updating these release notes...
- November 19, 2012: Added read_timeout option. (Feature only supported in standalone mode, will be supported by phpredis when pull request #260 is merged)
- October 29, 2012: Added support for persistent connections. (Thanks samm-git!)
- October 12, 2012: Improved memory usage and efficiency of garbage collection and updated recommendation.
- September 17, 2012: Added connect_retries option (default: 1) to prevent errors from random connection failures.
- July 10, 2012: Added password authentication support.
- Mar 1, 2012: Using latest Credis_Client which adds auto-reconnect for standalone mode.
- Feb 15, 2012: Changed from using separate keys for data, tags and mtime to a single hash per key.
- Nov 10, 2011: Changed from using phpredis and redisent to Credis (which wraps phpredis). Implemented pipelining.
@copyright Copyright (c) 2012 Colin Mollenhour (http://colin.mollenhour.com)
This project is licensed under the "New BSD" license (see source).