Showing posts with label Cloud Computing Performance. Show all posts
Showing posts with label Cloud Computing Performance. Show all posts

Friday, February 03, 2012

Uptime Report for my Live OakLeaf Systems Azure Table Services Sample Project: January 2012

My live OakLeaf Systems Azure Table Services Sample Project demo runs two small Windows Azure Web role instances from Microsoft’s South Central US (San Antonio, TX) data center. Here’s its uptime report from Pingdom.com for January 2012:

image


Following is detailed Pingdom response time data for the month of January 2012:

image


This is the eighth uptime report for the two-Web role version of the sample project. Reports will continue on a monthly basis.

Month Year Uptime Downtime Outages Response Time
January 2012 100.00% 00:00:00 0 773 ms
December 2011 100.00% 00:00:00 0 765 ms
November 2011 99.99% 00:05:00 1 708 ms
October 2011 99.99% 00:04:59 1 720 ms
September 2011 99.99% 00:05:00 1 743 ms
August 2011 99.98% 00:09:57 2 687 ms
July 2011 100.00% 00:00:00 0 643 ms
June 2011 100.00% 00:00:00 0 696 ms

The Azure Table Services Sample Project

See my Republished My Live Azure Table Storage Paging Demo App with Two Small Instances and Connect, RDP post of 5/9/2011 for more details of the Windows Azure test harness instance.

image_thumb2_thumb_thumb1_thumb

I believe this project is the oldest continuously running Windows Azure application. I first deployed it in November 2008 when I was writing Cloud Computing with the Windows Azure Platform for Wiley/WROX, which was the first book published about Windows Azure.

The application runs the default set of Windows Azure Diagnostics. About 5 GB of Event Counter data provide the source data for my Generating Big Data for Use with SQL Azure Federations and Apache Hadoop on Windows Azure Clusters post of 1/8/2012.

For more details about the Table Services Sample project or to download its source code, visit its Microsoft Pinpoint entry. Following are Pinpoint listings for three other related OakLeaf sample projects, two of which are live in the South Central US data center:

Monday, July 04, 2011

Video Keynote and Lightning Demo Archives from O’Reilly Media’s Velocity 2011 Conference, June 22-24, 2011

The Veloimagecity 2011 Web Performance & Operations Conference conference was held 6/22-6/24/2011 at the Santa Clara Convention Center in Santa Clara, CA. O’Reilly Media recorded all Velocity Conference 2011 keynotes, breakout sessions and workshops but only made the following video segments of keynote and lightning presentations available for public consumption:

Rachel Chalmers, "State of the Infrastructure " (14:52)

image It’s easy to forget that the story of infrastructure is a human story. In this session, we will take a step back and trace the history of automation and virtualization, touching on specific pioneers and the use cases they developed for both technologies, looking at the emergence of the devops role and its significance as organizations move towards the cloud. State of the Infrastructure Presentation [ZIP].

Rachel has led the infrastructure software practice for The 451 Group since its debut in April 2000. She pioneered coverage on SOA, distributed application management, utility computing and open source software. Today she focuses on datacenter automation and server, desktop and application virtualization.


Tim O'Reilly, "O'Reilly Radar" (10:54)

image Tim O’Reilly shares his insights into the world of emerging technology, presenting his take on what matters most – and what will be most disruptive – to the tech community.

Tim O’Reilly is the founder and CEO of O’Reilly Media, Inc. O’Reilly Media also hosts conferences on technology topics, including the O’Reilly Open Source Convention, the Web 2.0 Summit, and the Gov 2.0 Summit.


John Rauser, "Look at Your Data" (17:51)

image John has been extracting value from large datasets for 15 years at companies ranging from hedge funds and small data-driven startups to retailing giant amazon.com. He has deep experience in the areas of personalization, business intelligence, website performance and real-time fault analysis. An empiricist at heart, John’s optimism and can-do attitude make “Just do the experiment!” his favorite call to arms.

John has been extracting value from large datasets for 15 years at companies ranging from hedge funds and small data-driven startups to retailing giant amazon.com. He has deep experience in the areas of personalization, business intelligence, website performance and real-time fault analysis.


Douglas Crockford, "JavaScript & Metaperformance" (15:40)

image There are lies, damned lies, and benchmarks. Tuning language processor performance to benchmarks can have the unintended consequence of encouraging bad programming practices. This is the true story of developing a new benchmark with the intention of encouraging good practices.

Douglas Crockford is an Architect at Yahoo! Inc. He discovered JSON while he was CTO of State Software.


Theo Schlossnagle, "Career Development" (12:45)

image In this session, we will discuss how operations teams can extract key information from user-perceived performance measurements in real-time to make key operational assessments and decisions. Network routing collapses, CDN nodes fail, DNS Anycast horizons can shift and it can all be shown to you by your users. However, often times, this information isn’t correctly segmented and aggregated and the gems remain undiscovered. By looking at real-time user performance data and adding a good deal of magic meta information, we can now assess uncover serious operational problems before they subtly manifest is business reports later.

Through Javascript, browser web performance information and a good deal of real-time data analysis, you will dive to a deeper operational understanding of how your site is used by the world. Your NOC will be enlightened and forever better.

Theo Schlossnagle is a Founder and Principal at OmniTI where he designs and implements scalable solutions for highly trafficked sites and other clients in need of sound, scalable architectural engineering.


Mark Burgess, "Change = Mass x Velocity, and Other Laws of Infrastructure” (17:35)

image The key challenges for infrastructure designers and maintainers today are scale, speed and complexity. Mark Burgess was one of the first people to look for ways of managing these issues based on theoretical analysis. Much of his work has gone into the highly successful software Cfengine, which is still very much a leading light in the industry. In this session, Mark will ask if we have yet learned the lessons of infrastructure management, and, either way, what must come next.

Mark Burgess is the founder, CTO and principal author of Cfengine. He is Professor of Network and System Administration at Oslo University College and has led the way in theory and practice of automation and policy based management for 20 years.


Vik Chaudhary, "Testing and Monitoring the Smartphone Experience” (18:25)

image Keynote demonstrates how you can improve the end-user experience of your latest smartphone apps. This year at Velocity, Keynote debuts Mobile Device Perspective 5.0 (MDP), a cloud based testing and monitoring platform for ensuring the end to end quality of iPhone, Android and BlackBerry mobile apps accessing online content, streaming video, music and games. Learn how MDP’s mobile monitoring capabilities are designed for testing and optimizing smartphone access with gestures and touch events, using real mobile devices connected to the latest 3G and 4G networks in multiple mobile markets across the globe.

Mr. Chaudhary is responsible for leading Keynote’s product management team and has extended the company into new markets via 15 acquisitions.


Steve Souders, Jesse Robbins & John Allspaw, "Opening Remarks" (9:52)

image Steve works at Google on web performance and open source initiatives. He previously served as Chief Performance Yahoo!. Steve is the author of High Performance Web Sites and Even Faster Web Sites.

image Jesse Robbins (@jesserobbins) is CEO of Opscode and a widely recognized expert in Infrastructure, Web Operations, and Emergency Management. Jesse serves as co-chair of the Velocity Web Performance & Operations Conference and contributes to the O’Reilly Radar.

imageJohn has worked in systems operations for over fourteen years in biotech, government and online media. He built the backing infrastructures at Salon, InfoWorld, Friendster, and Flickr. He is now VP of Tech Operations at Etsy, and is the author of The Art of Capacity Planning published by O’Reilly.


Lew Tucker, "Cisco and Open Stack" (12:39)

image Cloud computing is changing the way we think about data centers, applications, and services. By pooling resources and aggregating workloads across multiple customers, cloud service providers achieve highly cost effective, multi-tenant, infrastructure as a service. Some applications, however, need a richer set of networking capabilities than are present in today’s environment. We will cover Cisco’s recent participation in the Open Stack community and explore what a networking service might provide for both application developers and service providers.

Lew Tucker is the Vice President and Chief Technology Officer of Cloud Computing at Cisco, where he is responsible for helping to shape the future of cloud and enterprise software strategies.


Jonathan Heiliger, "Facebook Open Compute & Other Infrastructure” (26:48)

image Jonathan Heiliger is the Vice President of Technical Operations at Facebook, where he oversees global infrastructure, site architecture and IT. Prior to Facebook, he was a technology advisor to several early-stage companies in connection with Index Ventures and Sequoia Capital. He formerly led the engineering team at Walmart.com, where he was responsible for infrastructure and building scalable systems. Jonathan also spent several years at Loudcloud (which became Opsware and was later acquired by HP) as the Chief Operating Officer.


John Resig, "Holistic Performance" (25:32)

image Working on the development of jQuery one tends to learn about all the performance implications of a particular change to a JavaScript code base (whether it be from an API change or a larger internals rewrite). Performance is an ever-present concern for every single commit and for every release. Performance implications must be well-defended and well-tested. In this talk we’re going to look at all the different performance concerns that the project deals with (processor, memory, network) and the tools that are used to make sure development continues to move smoothly.

John Resig is a JavaScript Evangelist for the Mozilla Corporation and the author of the book Pro JavaScript Techniques. He’s also the creator and lead developer of the jQuery JavaScript library.


Paul Querna, "Cast - The Open Deployment Platform" (4:10)

image Cast is an open source platform for deploying and managing applications being developed by Rackspace. Cast is the bridge between traditional configuration management like Puppet or Chef, and the longer term shift to Platforms as a Service. Cast is written in Node.js, built entirely on top of HTTP APIs for easy control, and has plugins and hooks to make customization trivial.

Paul Querna is an Architect at Rackspace, and former Chief Architect at Cloudkick.


Jon Jenkins, "Velocity Culture" (15:14)

image Operations is can play a critical role in driving revenue for the business. This talk will explore some ways in which the ops team at Amazon is thinking outside the box to drive profitability. JJ will also issue a challenge for next year’s Velocity Conference.

Jon Jenkins is the Director of Platform Analysis at Amazon.com. He leads a team focused on designing and implementing highly-available architectures capable of operating within tight performance requirements at massive scale with minimal human intervention.


Michael Kuperman & Ronni Zehavi, "Your Mobile Performance – Analyze and Accelerate” (4:45)

image Michael Kuperman has 10 years of experience in the CDN and ADN industry. To his role as VP of Operations at Cotendo he brings deep expertise in deploying and operating mission critical global environments in the US, Europe and Asia. Before joining Cotendo, Michael held senior Operations positions at a number of leading companies in the content delivery network industry.

image Ronni Zehavi is CEO and Co-Founder of Cotendo. He has over fifteen years of executive experience in start-up and multi-national high-technology companies. Prior to Cotendo he founded Commtouch Software where he commercialized email security technologies in a distributed global service environment that handled over a billion content transactions per day with 100% uptime.


Patrick Lightbody, "From Inception to Acquisition" (15:29)

image Launched in December 2008, BrowserMob set out to change the way performance and load testing is done – all using the cloud. Learn from the founder how he built a high performance testing product on top of dozens of cloud technologies, and how the operational support the cloud provided enabled the company to not only profit from day one, but to be acquired within a year and a half of its launch.

Patrick Lightbody is the founder of BrowserMob, a cutting edge load testing and website monitoring provider, which is now a part of Neustar. He also founded HostedQA, an automated web testing platform.


Velocity 2011: Ian Flint, "World IPv6 Day: What We Learned" (18:24)

image One of the greatest challenges facing an organization that wants to migrate to IPv6 is that the protocol is not supported on all clients. In fact, it is estimated that 1/2000 of clients will be unable to connect to a dual-stack site.

This poses an interesting question; who goes first? Enabling IPv6 in the current climate means sending customers to the competition. But if no one goes first, IPv6 will never happen.

This is the motivating factor behind World IPv6 Day. Major content providers have agreed to simultaneously enable IPv6 for a 24-hour period in early June. What this means is that rather than one site seeming broken, the whole Internet will seem broken to clients with this problem, which will call attention to deficiencies in the infrastructure of the Internet, rather than to individual content providers.

Enabling IPv6 is a full stack exercise – everything from routers to applications needs to be adapted to work with the new protocols. The purpose of this session is to share lessons learned in the preparation for World IPv6 day, and discoveries made on the day itself.

Ian Flint is a 14-year veteran of the Internet. He has founded and sold two successful startups, and has been an architect at eBay and Yahoo. He is currently the service architect for communities and communications at Yahoo.


Arvind Jain, "Making the Web Instant" (13:40)

image Since last August, you have been able to get Google Search results instantly, even before you are done typing your query. But, that’s not enough. When you click on a result, you want that page to load instantly as well. If you don’t search, but navigate directly to a web page from the address bar, that should load instantly too.

We have implemented prerendering in Chrome to make these possible. This experimental feature utilizes the time when you are thinking or typing to predict what you need next, and gets it ready for you ahead of time. It is a challenging problem. A wrong prediction makes ads and analytics accounting inaccurate, wastes client resources and increases server load. We discuss how we address these concerns

Arvind Jain works as a Director of Enginering at Google. He is responsible for making Google products fast and also runs its “Make the Web faster” initiative.


Artur Bergman, "Artur on SSD's" (3:55)

image Artur Bergman, hacker and technologist at-large, is the VP of Engineering and Operations at Wikia. He provides the technical backbone necessary for Wikia’s mission to compile and index the world’s knowledge. … In past lives, he’s built high volume financial trading systems, re-implemented Perl 5’s threading system, wrote djabberd, managed LiveJournal’s engineering team, and served as operations architect at Six Apart.


"Lightning Demos Wednesday" (30:46)

"Lightning Demos Thursday" (33:12)


Tim O’Reilly is a strong proponent of free and open-source software but he doesn’t believe conference breakout session and workshop videos should be free and open. You must purchase a $495 online access pass to view the entire conference video archive.

All session videos from later DevOps Days Mountain View 2011 and Giga Om’s Structure 2011 conferences are available to view at no charge. For Structure 2011 video archives, see my Giga Om Structure Conference 2011 - Links to Archived Videos for 6/22/2011 and Giga Om Structure Conference 2011 - Links to Archived Videos for 6/23/2011 posts.


Monday, May 09, 2011

Republished My Live Azure Table Storage Paging Demo App with Two Small Instances and Connect, RDP

Updated 5/9/2011: Changed from ExtraSmall to Small instances because MSDN Ultimate subscribers and MSDN Premium subscribers whose benefit has been upgraded to Ultimate have a free 1,500 hours of Small compute instances, not ExtraSmall instances. See my Microsoft Bills for ExtraSmall Instances with a MSDN Ultimate Benefit post of 5/9/2011 for more details.


On 2/4/2010, Steve Marx (@smarx) asked me to stop publishing monthly uptime reports for a single instance of my live OakLeaf Systems Azure Table Services Sample Project - Paging and Batch Updates Demo because they were irrelevant, misleading and not representative of ordinary production deployments. Steve likened my approach to reporting the uptime of a bank by observing the presence or absence of a single teller. I acquiesced and stopped the monthly reports. My Windows Azure Uptime Report: OakLeaf Table Test Harness for January 2011 (100.00%), which contains Steve’s request in a comment, was the last.

Microsoft requires two instances of Web or Worker roles to qualify for the Windows Azure Compute Service Level Agreement (SLA):

Windows Azure has separate SLA’s for compute and storage. For compute, we guarantee that when you deploy two or more role instances in different fault and upgrade domains your Internet facing roles will have external connectivity at least 99.95% of the time. Additionally, we will monitor all of your individual role instances and guarantee that 99.9% of the time we will detect when a role instance’s process is not running and initiate corrective action.

The separate Windows Azure Storage SLA is as follows:

For storage, we guarantee that at least 99.9% of the time we will successfully process correctly formatted requests that we receive to add, update, read and delete data. We also guarantee that your storage accounts will have connectivity to our Internet gateway.

Click here for summaries of all Windows Azure SLAs.

My MSDN Ultimate subscription benefit changed from 750 hours/month of a Small instance to 1,500 hours/month of ExtraSmall instances in April 2011, which enabled running two simultaneous compute instances without charging my credit card. On 5/8/2011, I upgraded my sample project as follows:

  • Changed from a single Small compute instance to two ExtraSmall compute instances
  • Added IntelliSense for deployment troubleshooting
  • Added Windows a Azure Connect VPN and enabled Remote Desktop Access
  • Updated from Windows Azure SDK v1.2 to v1.4

image

I had enabled enabled diagnostics on 12/4/2011, as reported in my Adding Trace, Event, Counter and Error Logging to the OakLeaf Systems Azure Table Services Sample Project post of 12/5/2011.

I wasn’t able to perform a Swap VIP operation because my existing production instance had only one Web Role and the new site would have two. Therefore, I had to delete the existing production Web Role and publish the project with Visual Studio 2010 to a new Production site.

Deployment Speed Record: < 11 Minutes

Deploying the new site on Sunday morning established a new speed record for me: Less than 11 minutes from start to finish. Here’s the Visual Studio Activity Log:

7:08:14 AM - Preparing...
7:08:14 AM - Connecting...
7:08:15 AM - Uploading...
7:10:46 AM - Creating...
7:11:22 AM - Starting...
7:11:59 AM - Initializing...
7:11:59 AM - Instance 0 of role WebRole1 is initializing
7:11:59 AM - Instance 1 of role WebRole1 is initializing
7:16:51 AM - Instance 0 of role WebRole1 is busy
7:16:51 AM - Instance 1 of role WebRole1 is busy
7:17:56 AM - Instance 1 of role WebRole1 is ready
7:19:01 AM - Instance 0 of role WebRole1 is ready
7:19:01 AM - Complete.

Update 5/9/2011: The Small-instance replacement took about 00:11:15 on Monday morning, so the earlier provisioning time wasn’t a fluke.

Contrast this time with the 32 minutes it took me to activate one IBM SmartCloud Enterprise Silver instance, as reported in the Connecting to the instance with Windows' Remote Desktop Protocol section of my Test-Driving IBM’s SmartCloud Enterprise Infrastructure as a Service: Part 2 - Spring 2011 Promotion Free Trial, updated 5/6/2011.

Here’s the Windows Azure Portal page for the upgraded site:

image

Note that the Quota’s maximum Cores (20) applies to paid subscriptions. Only 2 cores are free with MSDN Premium and Ultimate benefits.

I plan to resume publishing uptime reports for oakleaf.cloudapp.net based on Pingdom availability data beginning the first week of July for June 2011.

For more details about the test harness and its qualification to use Microsoft Platform Ready (MPR) program’s “Powered by Windows Azure” logo, see my Azure Storage Services Test Harness: Table Services 1 – Introduction and Overview post (updated 11/17/2010).

Monday, April 04, 2011

Making Sense of Window Azure’s Global Peformance Ratings by CloudSleuth

Update 4/4/2011 12:30 PM PDT: Ryan Bateman of Compuware Gomez/CloudSleuth responded to this post on Twitter. See end of article for tweets.

Lori MacVittie (@lmacvittie) asserted Application performance is more and more about dependencies in the delivery chain, not the application itself in a preface to her On Cloud, Integration and Performance post of 4/4/2011 to F5’s DevCentral blog:

image When an article regarding cloud performance and an associated average $1M in loss as a result appeared it caused an uproar in the Twittersphere, at least amongst the Clouderati. There was much gnashing of teeth and pounding of fists that ultimately led to questioning the methodology and ultimately the veracity of the report.

quote-badgeIf you were worried about the performance of cloud-based applications, here's fair warning: You'll probably be even more so when you consider findings from a recent survey conducted by Vanson Bourne on the behalf of Compuware.

On average, IT directors at 378 large enterprises in North America reported their organizations lost almost $1 million annually due to poorly performing cloud-based applications. More specifically, close to one-third of these companies are losing $1.5 million or more per year, with 14% reporting that losses reach $3 million or more a year.

In a newly released white paper, called "Performance in the Cloud," Compuware notes its previous research showing that 33% of users will abandon a page and go elsewhere when response times reach six seconds. Response times are a notable worry, then, when the application rides over the complex and extended delivery chain that is cloud, the company says.

-- The cost of bad cloud-based application performance, Beth Shultz, NetworkWorld 

I had a chance to chat with Compuware about the survey after the hoopla and dug up some interesting tidbits – all of which were absolutely valid points and concerns regarding the nature of performance and cloud-based applications.

image What the Compuware performance survey does is highlight the very real problem with measuring that performance from a provider point of view. It’s one thing to suggest that IT find a way to measure applications holistically and application performance vendors like Compuware will be quick to point out that agents are more than capable of not only measuring the performance of individual services comprising an application but that’s only part of the performance picture. As we grow increasingly dependent on third-party, off-premise and cloud-based services for application functionality and business processing we will need to find a better way to integrate performance monitoring into IT as well. And therein lies the biggest challenge of a hyper-connected, distributed application. Without some kind of standardized measurement and monitoring services for those application and business related services, there’s no consistency in measurement across customers. No measurement means no visibility, and no visibility means a more challenging chore for IT operations to optimize, manage, and provision appropriately in the face of degrading performance.

Application performance monitoring and management doesn’t scale well in the face of off-premise distributed third-party provided services. Cloud-based applications IT deploys and controls can employ agents or other enterprise-standard monitoring and management as part of the deployment, but they have no visibility into let alone control over Twitter or their supply-chain provider’s services.

It’s a challenge that will continue to plague IT for the foreseeable future, until some method of providing visibility into those services, at least, is provided such that IT and operations can make the appropriate adjustments (compensatory controls) internal to the data center to address any performance issues arising from the use of third-party provided services.

CompuWare’s CloudSleuth application is based on the company’s Gomez SaaS-based application performance management (APM) solution claimed to “provide visibility across the entire web application delivery chain, from the First Mile (data center) to the Last Mile (end user).”


Charles Babcock claimed “CloudSleuth's comparison of 13 cloud services showed a tight race among Microsoft, Google App Engine, GoGrid, Amazon EC2, and Rackspace” in a deck for his Microsoft Azure Named Fastest Cloud Service article of 3/4/2011:

image In a comparative measure of cloud service providers, Microsoft's Windows Azure has come out ahead. Azure offered the fastest response times to end users for a standard e-commerce application. But the amount of time that separated the top five public cloud vendors was minuscule.

imageThese are first results I know of that try to show the ability of various providers to deliver a workload result. The same application was placed in each vendor's cloud, then banged on by thousands of automated users over the course of 11 months. …

imageGoogle App Engine was the number two service, followed by GoGrid, market leader Amazon EC2, and Rackspace, according to tests by Compuware's CloudSleuth service. The top five were all within 0.8 second of each other, indicating the top service providers show a similar ability to deliver responses from a transaction application.

imageFor example, the test involved the ability to deliver a Web page filled with catalog-type information consisting of many small images and text details, followed by a second page consisting of a large image and labels. Microsoft's Azure cloud data center outside Chicago was able to execute the required steps in 10.142 seconds. GoGrid delivered results in 10.468 seconds and Amazon's EC2 Northern Virginia data center weighed in at 10.942 seconds. Rackspace delivered in 10.999 seconds.

It's amazing, given the variety of architectures and management techniques involved, that the top five show such similar results. "Those guys are all doing a great job," said Doug Willoughby, director of cloud strategy at Compuware. I tend to agree. The results listed are averages for the month of December, when traffic increased at many providers. Results for October and November were slightly lower, between 9 and 10 seconds.

The response times might seem long compared to, say, the Google search engine's less than a second responses. But the test application is designed to require a multi-step transaction that's being requested by users from a variety of locations around the world. CloudSleuth launches queries to the application from an agent placed on 150,000 user computers.

It's the largest bot network in the world, said Willoughby, then he corrected himself. "It's the largest legal bot network," since malware bot networks of considerable size keep springing up from time to time.

Page 2:  Testing Reflects The Last Mile

Read more: 2, Next Page »


Ryan Bateman (@ryanbateman) ranked Windows Azure #1 of the top 15 cloud service providers in his Cloud Provider Global Performance Ranking – January post of 3/16/2011 to the CloudSleuth blog (missed when posted):

image A few weeks ago we released calendar year Q4 “Top 15 Cloud Service Providers – Ranked by Global Performance”.  My intention here is to start a trend; releasing our ranking of the top cloud service providers based on a global average of real-time performance results as seen from the Gomez Performance Network and its Last Mile nodes.

image

If you missed out on the last post and are curious about the methodology behind these numbers, here is the gist:

image1. We (Compuware) signed up for services with each of the cloud providers above.

2. We provisioned an identical sample application to each provider.  This application represents a simple ecommerce design, one page with sample text, thumbnails and generic nav functions followed by another similar page with a larger image.  This generic app is specifically designed as not to favor the performance strengths of any one provider.

3. We choose 200 of the 150,000+ Gomez Last Mile peers (real user PCs) each hour to run performance tests from.  Of the 200 peers selected, 125 are in the US.  The remaining 75 are spread across the top 30 countries based on GDP.

4. We gather those results and make them available through the Global Provider View.

These decisions are fueled by our desire to deliver the closest thing to an apples-to-apples comparison of the performance of these cloud providers.  Got ideas about how to tweak the sample application or shift the Last Mile peer blend or anything else to make this effort more accurate?  Let us know, we take your suggestions seriously.


Ryan’s post was reported (but usually not linked) by many industry pundits and bloggers.

Click here for more details about the technology Cloudsleuth uses.

Here’s a 4/1/2011 screen capture of the worldwide response-time results for the last 30 days with Google App Engine back in first place:

image

Note that the weighting of US users causes results from Amazon EC2 (EU – Ireland), Amazon EC2 (APAC – Tokyo) and Windows Azure (Southeast Asia) to be much slower than what users can expect in regions closer to these data centers.

image

Windows Azure still takes the top spot in North America, as shown here:

image

Checking response time by city by clicking the arrow at the right of the Provider column indicates that the data is for Microsoft’s North Central US (Chicago) data center:

4744image

The following Global Provider View of 4/4/2011 indicates that the Global App Engine is located near Washington, DC (probably in Northern Virginia):

image

Unlike Amazon Web Services, for which CloudSleuth returns results for four regional data centers (US East – Northern Virginia, US West – Northern California, EU – Ireland, and APAC – Tokyo), data for Windows Azure is limited to two (US – North Central and Southeast Asia [Singapore]) data centers and Google App Engine only one.

Collecting data based on distant data centers when the provider has closer data centers unfairly penalizes those with multinational coverage, as shown here on 4/4/2011:

image

Chicago, IL is 587 miles further from Europe than Washington, DC, which could account for at least some of the difference in performance between Google App Engine and Windows Azure. Microsoft has European data centers for Windows Azure in Amsterdam and Dublin.

Google discloses public Data Center Locations on a Corporate page with the following details:

Click on the link for a specific location to read more about our data center there, as well as find out about available jobs and learn about our community efforts in the region.

United States
Europe

Conclusion:

CloudSleuth should revise its performance measurement methodology to use the provider’s data center that’s closest to the region of interest.

Note: Major parts of this post previously appeared in my Windows Azure and Cloud Computing Posts for 3/4/2011+ and Windows Azure and Cloud Computing Posts for 3/30/2011+.


Update 4/4/2011 12:30 PM PDT: Ryan Bateman tweeted the following:

image

I replied with this question:

image 

Waiting for Ryan’s response. …