Sponsored by LoadStorm. The easy and cost effective load testing tool for web and mobile applications.
The full company list, product list, methodology, and notes can be found here
This post focuses only on the results of the testing in the <$25/month price bracket for WordPress Hosting.
Other Price Tier Results
<$25/Month Tier$25-50/Month Tier$51-100/Month Tier$101-200/Month Tier$201-500/Month Tier$500+/Month (Enterprise) Tier
<$25/Month WordPress Hosting Products
<$25/Month WordPress Hosting Performance Benchmarks Results
1. Load Storm
Test 500-2000 Concurrent Users over 30 Minutes, 10 Minutes at Peak
|Company||Total Requests||Total Errors||Peak RPS||Average RPS||Peak Response Time(ms)||Average Response Time(ms)||Total Data Transferred (GB)||Peak Throughput (MB/s)||Average Throughput (MB/s)|
|Hosting Agency (DE)||182424||117939||132.65||101.35||15991||6743||3.823||10.53||2.124|
Discussion of Load Storm Test Results
The companies that clearly didn't struggle at all with LoadStorm were DreamHost [Reviews], Incendia Web Works (IWW), LightningBase, Pressed, SiteGround [Reviews]. GoDaddy [Reviews], MediaTemple [Reviews] and Traffic Planet Hosting had minor spikes at the start, but they seem nearly inconsequential in the grand scheme of the test.
WP.land seemed to have some security measures which struggled with wp-login being hit so frequently.
A2 Hosting [Reviews], BlueHost [Reviews], FlyWheel [Reviews] and Hosting Agency did not do well on this test. FlyWheel explicitly stated this was too much load for that size plan and recommended upgrading if this was the expected load.
Test 1-1000 Concurrent Users over 60 seconds
Blitz Test Quick Results Table
|Company||Hits||Errors||Timeouts||Average Hits/Second||Average Response Time||Fastest Response||Slowest Response|
|Hosting Agency (DE)||662||20862||3649||11||630||400||1556|
Discussion of Blitz Test 1 Results
This test is just testing whether the company is caching the front page and how well whatever caching system they have setup is performing (generally this hits something like Varnish or Nginx).
Who performed without any major issues?
DreamHost, IWW, LightningBase, SiteGround, WP Land all handled the test without any issues.
Who had some minor issues?
BlueHost had a couple spikes during the test which caused some errors and timeouts, but they weren't substantial.
FlyWheel had a spike at the very end of the test which caused a large increase in response times.
Pressed started to have a ramp up in response times but it never errored or timed out during the test.
Who had some major issues?
GoDaddy, MediaTemple and TrafficPlanetHosting seemed to pretty clearly hit security measures which couldn't be worked around. The response times were relatively stable, but errors shot up which is symptomatic of a security measure kicking in rather than the server being taxed. It's hard to know how they would have performed sans security measures.
A2 and Hosting Agency did not take kindly to the Blitz test and crashed almost immediately under load.
3. Uptime Monitoring
Both uptime monitoring solutions were third party providers that offer free services. UptimeRobot was paid for and monitoring on a 1 minute interval. All the companies were monitored over approximately two months (May-June 2016).
|Hosting Agency (DE)||-||100|
BlueHost screwed up and cancelled this account mid-testing causing the uptime to look horrific. Their other two plans which were not cancelled had measurements of 99.98, 99.98, 100 and 99.99 uptime. I'm upset that it happened and there was a struggle to restore the account and have to take credit away for this type of screw up. But, they were able to keep the other servers up with near perfect uptime which I think should be stated here as well.
Hosting Agency for some reason couldn't be monitored by StatusCake (http/2 issue they still haven't fixed for nearly 9 months, which UptimeRobot fixed within 24 hours when I notified them). But they had 100% on UptimeRobot, so it looks good.
IWW had a bunch of short outages and one longer one (2hr 33m) which brought it's uptime down.
Pressed had a 1hr 51m downtime (502 error) recorded by UptimeRobot but StatusCake never picked it up. I'm not sure what to make of that, it might be something wrong with UptimeRobot's servers connecting properly since StatusCake never picked it up over an interval that long.
Everyone else had above 99.9% uptime.
Every test was run with the settings: Chrome Browser, 9 Runs, native connection (no traffic shaping), first view only.
|Company||WPT Dulles||WPT Denver||WPT LA||WPT London||WPT Frankfurt||WPT South Africa|
|Hosting Agency (DE)||1.299||1.258||2.17||0.985||1.55||4.905|
|Company||WPT Singapore||WPT Shanghai||WPT Japan||WPT Sydney||WPT Brazil|
|Hosting Agency (DE)||2.311||22.406||2.651||2.772||2.596|
What I learned was getting traffic into China is terrible. Nobody really did well on the Shanghai location. South Africa is also really slow. Most servers were US based but were delivering content to most corners of the world in about 2 seconds or less which is impressive. Hosting Agency based in Germany was a bit disappointing. Very slow relatively speaking to the US. But it wasn't even the fastest to London or Frankfurt. LightningBase and IWW were able to beat the German company in the US by a large margin and to Europe which reinforces that geographic location isn't everything in terms of speed.
I wish I could compare averages against last year except they removed one of the testing locations (Miami) and I did a global test instead because that was something people wanted to see.
|Company||PHP Bench [Seconds] (lower=faster)||WP Bench [Queries Per Second](higher=faster)|
|Hosting Agency (DE)||6.501||45.28|
What was enormously interesting about WPPerformanceTester results this year was the much larger spread and faster results. Last year, almost everyone was around 10-14 seconds for PHP Bench with the outlier of PressLabs doing 8.9 and DreamHost at 27. DreamHost again has the dubious honor of the slowest PHP Bench but it improved by a whopping 10 seconds down to 17. The fastest was Hosting Agency with 6.5, more than a full 2 seconds faster than last year's fastest speed. IWW, TrafficPlanetHosting also managed sub 10 second speeds.
Last year's fastest WP Bench was 889 queries per second. That was blown away by this years testing with IWW leading the group at more than double the speed (1869). BlueHost, LightningBase and TrafficPlanetHosting all managed to be faster than last year's fastest benchmark as well. Unfortunately, Hosting Agency's incredibly fast PHP bench is somewhat cancelled out by their slowest WP Bench score, which is slower than last year's slowest. It should be noted that transaction speed isn't always a great measured on distributed/clustered/cloud systems that may be running databases on different machines, but at the entry level that's less of an issue. Generally the incredibly fast scores you see are local databases with no network latency overhead.
It is nice to get back to a real entry level analysis with a much more level playing field. Having 13 different companies available to choose from in the <$25/month range is fantastic. Despite the change in this years format, the lower end plans still outperformed the fastest competitors from last year's tests which had plans up to ~$300/month.
Despite the hard price cap in this bracket of testing, there were still some companies that handled all the tests without any serious issue. Many more did very well but ran into minor issues.
The amount of companies jumping into the space is a fantastic win for consumers. In this tier we saw A2, Pressed, WP Land, Hosting Agency, IWW and Traffic Planet Hosting all enter for the first time. They target a variety of different niches within the space and overall it's a win for us, the consumer to have more good choices and options. From a performance standpoint, you can still get amazing performance value for the money even at the lowest tier.
Without further ado, I will tell you who had the best performance, who deserved an honorable mention and then analyze each host individually. I still don't believe in ranking in any particular order, only grouping companies by how well they performed.
Top Tier WordPress Hosting Performance
DreamHost [Reviews], LightningBase, and SiteGround [Reviews],
All three of these companies went through the full testing without any meaningful issues.
Pressed had an odd uptime issue but also showed some signs of server stress during the blitz test. For a brand new company they performed admirably, but I'm not quite comfortable awarding them the top tier status quite yet when you compare their results against the three top tier companies, but they put on a very good showing.
WP.land did well in every test except LoadStorm where it had a roughly 4% error rate. It looked like a security issue with wp-login which isn't uncommon. But there were also some spikes/delays as well. It could just be security acting up, but again, a minor issue that kept it out of the top tier, but it was worthy of an honorable mention from yet another new comer to this year's testing.
GoDaddy [Reviews]/MediaTemple [Reviews], I combine this one because it's running on the same tech and the results look very similar and experienced the same security issues. You can pretty clearly see when the security measures kick in on Blitz and I wasn't able to work with their tech team to come up with a way to responsibly bypass their security measures. LoadStorm had a spike at the start with wp-login issues but resolved itself out quickly and had a flat response time graph. It's possible their tech is just as good as the top tier hosts, but I wasn't able to accurately measure it because of security measures but it looks very good and at least deserves the honorable mention.
Traffic Planet Hosting is another new entrant and had similar issues to GoDaddy/MediaTemple. Security issues caused some problems on the Blitz test, but it did start to show some load too. Not perfect, but it did well on LoadStorm as well. (no honorable mention?)
Individual Host Analysis
A2 Hosting was a new entrant to this test and as much as I love the competition in the space, A2 fell short. Other than their uptime monitoring which was good, they struggled in all the load testing experiments.
BlueHost specifically messed up with my account in this test and the uptime was terrible because of it. That alone ruined the uptime test, although as I stated in the section, the other servers all maintained excellent uptime which were on different accounts. They did ok in the blitz test, but not in the LoadStorm test. They also surprisingly managed the fastest individual WebPageTest score of any host in this price range. Compared to last year I don't see any huge signs of improvement with regards to performance.
Last year DreamHost's DreamPress product almost made the top tier except for some major downtime issues. This year, they had no such downtime issues and the performance remained top notch. DreamHost earned the top tier status for the <$25/month price bracket. It appears to be an excellent product priced very competitively.
FlyWheel only entered one product this year and it was less powerful than last year's. It struggled a bit more on the LoadStorm test but the Blitz was perfect (although for this price tier, it was a weaker test than last year's test). They explicitly stated for LoadStorm that the plan was inappropriate for that level of traffic. They can probably handle bigger sites, but if we're comparing dollars to performance, they fell short in this price bracket on that metric. But they are still rated as the most well liked company that we track at Review Signal, so they are clearly doing something right in terms of product and customer service.
GoDaddy had a stalwart performance marred by what appeared to be security measures. They very well could have a top notch product but we couldn't work out a responsible way to bypass the security measures for the Blitz load test. LoadStorm looked pretty good, one small spike to start and steady up to 2000 users. GoDaddy earned an honorable mention status because the product didn't seem to encounter any non-artificial problems.
IWW did a great job in both load tests. The only concern was uptime, where IWW had 99.73% and 99.88% as recorded by each service. The performance component is definitely there, but a little more consistency and we have another serious competitor in the space. The only reason they didn't earn honorable mention while Pressed did is that there were conflicting uptime reports for Pressed where one showed 100% and the other recorded sub 99.9% uptime. Two independent services showed IWW below 99.9%, so there isn't much doubt about it in my mind. Like DreamHost last year, they put on a great performance showing and I hope next year the servers are a bit more stable and I can award top tier status.
LightningBase continues to impress. The last two years they've put on consistently near perfect tests. Their Blitz result was perfect and their LoadStorm had only 5 errors out of 314439 requests. Combined with 100/99.99% uptime monitors, LightningBase is unquestionably in the top tier for the <$25/month WordPress hosting bracket.
MediaTemple's results basically mirrored GoDaddy's results. It would be even hard to tell the graphs apart if you removed the names. The MediaTemple/GoDaddy platform appears to be very solid but we couldn't responsibly get by some security measures, so I couldn't award it top tier status, but MT earned an honorable mention.
Pressed earned itself an honorable mention. It had a weird uptime issue but more importantly it started to show some signs of load during the Blitz test where I would expect a flat response time from a static cache test like Blitz. It's a very new product and I'm sure we'll continue to see tremendous improvements as time goes on, a very good performance from possibly the newest company in this year's testing.
Hosting Agency performed as expected, it appears to have no special WordPress optimizations. If you were to install a basic lamp stack, this is the performance I expect out of the box. They had perfect uptime and oddly found themselves on both ends of the spectrum on my WPPerformanceTester. They weren't faster to England or Germany on WebPageTest, which I suspect is because there was no special caching technologies to accelerate delivery of pages despite being geographically closer. And it just collapsed during the load tests, especially Blitz which is essentially a static cache test (where they have none). Another important note is that their entire system is in German only.
SiteGround got even better this year. They jumped up from honorable mention to top tier status. Their Blitz and LoadStorm tests both improved while everything else remained at a high level. An all around fantastic performance which deserved top tier status.
Another new comer to this years testing. TPH put on a good show, there seemed to be some security measures which ruined the Blitz testing, but the LoadStorm test looked very solid. They earned an honorable mention because the only issue seemed artificial. I'm less confident about the quality of the product than GoDaddy/MediaTemple, but it still seemed to warrant recognition.
WPLand was the final new entrant and they put on a fantastic showing. Everything went near perfect except the LoadStorm test which seemed to have an issue with wp-login triggering some security measures. But the response rate was pretty stable and quick despite the ramp up to 2000 users. They also had a perfect blitz test with no errors and a 1ms spread in fastest to slowest response times. WP Land earned honorable mention status because overall it was a very good performance with a small issue that might be security related.
Other Price Tier Results
<$25/Month Tier$25-50/Month Tier$51-100/Month Tier$101-200/Month Tier$201-500/Month Tier$500+/Month (Enterprise) Tier
Latest posts by Kevin Ohashi (see all)
- Analyzing Digital Ocean’s First Major Move with Cloudways - February 28, 2023
- Removing old companies - June 28, 2021
- WordPress & WooCommerce Hosting Performance Benchmarks 2021 - May 27, 2021
Pingback: $51-100/Month WordPress Hosting Performance Benchmarks (2016) | Review Signal Blog
Pingback: $201-500/Month WordPress Hosting Performance Benchmarks (2016) | Review Signal Blog
Pingback: WordPress Hosting Performance Benchmarks (2016) | Review Signal Blog
Pingback: $25-50/Month WordPress Hosting Performance Benchmarks (2016) | Review Signal Blog
Pingback: $101-200/Month WordPress Hosting Performance Benchmarks (2016) | Review Signal Blog
Pingback: $500+/Month Enterprise WordPress Hosting Performance Benchmarks (2016) | Review Signal Blog
Any reason why Inmotion hosting was not considered?
Yes. They didn’t ask to be included. It’s an opt-in test.
I don’t know if they really do high performance WordPress or just sell hosting labeled as WordPress hosting. If its the latter, these tests will make a company look bad. It’s for companies that really specialize in WordPress including making it performant.
Is there a reason why Flywheel opted to leave their Personal, Professional and Custom plans out of the rest of the tests?
I don’t have an answer for that, you would have to ask them.
After reading your reviews last year I moved to LightingBase. Thank you so much for providing such useful information. You really helped me find the perfect host for me – their support is amazing too.
It makes me really happy that you’re happy with your host 🙂
What about DigitalOcean?
They didn’t compete and probably with good reason. These companies pretty much all do managed WordPress hosting. Digital Ocean is unmanaged VPS hosting where the user is expected to do it themselves. Many companies listed here are built on top of Digital Ocean infrastructure (FlyWheel, CloudWays jump to mind, maybe more).
What if the tested host combined with CloudFlare?
Would it affect the basic result performance?
It could improve it potentially depending on the test. The LoadStorm test it shouldn’t help as much because the users log in which has to hit the primary servers and bypass any CDN.
Have you ever considered 1and1 also as a shared hoster or wordpress hoster?
This testing is opt-in, 1&1 did not opt-in this year.
1&1 is trash, as is the vast majority of shared hosting (GoDaddy, HostGator, BlueHost, etc). You get what you pay for when you go with cheap hosting.
SiteGround is the only shared host that I’ve found that is any good (aka isn’t slow as hell, and rarely sees drastic load time reductions during peak traffic periods), and even then they’re only good if you use their $7.95 and up plans – their $3.95/mo plan is way too slow – the higher plans have caching that significantly speeds them up.
honestly, what do you expect for a $3.95/m plan?
It is the very entry level for those who just barely wants to be online and have just a few visitors a day.
Anyone who is serious about their website and want performance will be happy to pay $10-15 per months and for that you can even get a 1GB VPS nowadays which will blow any shared host out the water.
Siteground is great with pricing, performance and reliability. My first choice for WP.
I can’t express my gratitude enough for these benchmarks. Your site is really helpful as it provides REAL reviews, instead of those crap “top10” lists.
Seriously, thank you very much!
Keep up the good work and please, do another one in 2017.
Thanks for the detailed review, we have our development sites hosted over at pantheon. But, have you tried / considered fastcomet? They also have WordPress plans, and wanna see how well they’ve performed. Hoping you’ll consider testing them.
The tests are entirely opt-in, so if you know someone there that wants to participate, have them reach out to me.
Are your figures for BlueHost based on their generic shared hosting plan, or their WordPress hosting plan? I see much different results for BlueHost on some other performance reviews.
I don’t think the plan exists anymore, it was their Blogger WordPress hosting plan. I couldn’t find it on the site any longer though.
Why no hostgator?
They didn’t opt in. I need permission to do testing like this because it can potentially affect other customers and trigger security systems. BlueHost is owed by the same group though and was include. Not sure if the tech is the same or not though.