Welcome!

Microsoft Cloud Authors: David H Deans, Yeshim Deniz, Janakiram MSV, Andreas Grabner, Stackify Blog

Blog Feed Post

Deploying APM in the Enterprise Part 3 – Getting Started With APM

Welcome back to my series on Deploying APM in the Enterprise. In Part 2 we discussed APM maturity from an unusual perspective as maturity models go. In this blog we are going to discuss the process of getting what you want and need in the way of an APM product.

Warning – This is an astronomically long blog post so make sure you have plenty of time to read it through.

Let’s assume for a minute that you actually read Part 2 – APM Maturity As You’ve Probably Never Seen It Before and have an idea where you and your organization currently are on the maturity scale. To advance to a higher level of maturity you can work on process and people until the cows come home but eventually you are going to want a tool that provides insight your people and processes never will. So that is where I am going to focus the rest of this series while touching on the people and process aspect throughout.

If you have decided you need or want to buy a new software tool here is my tried and proven method for making it happen. Just to make things a little more interesting each step in the process is a song title that relates somehow and that I actually like. So feel free to comment on the blog post and my taste in music if you feel like it.

Throughout this blog post I will share some key lessons learned from my own experience dealing with software vendors. This information will be highlighted in bold red text so it is easier to pick out.

It Hurts So Good – John Cougar

Nobody will agree to spend money on a tool unless there is some problem putting the hurt on your business (lost revenue, productivity impact, customer satisfaction, etc…). Find and document a tangible problem. Preferably an issue with a business/mission critical application like your e-commerce platform, online trading, payment gateway, risk calculation, settlement system, etc… Find some application or service that is impacting your business in a meaningful way due to poor performance and/or downtime and document the following:

  • Number of issues and severity level
  • Mean Time To Repair (MTTR – usually the average amount of time from first impact to problem resolution)
  • Quantifiable measure of impact on business (dollars lost per minute, potential customers lost, trades lost per minute, etc…)
  • Average number of employees involved in troubleshooting each issue
  • Root cause of each incident

You will use this data in your evaluation document and your business justification down the road.

The World I Know – Collective Soul

“So I walk up on high / And I step to the edge / To see my world below” If you haven’t already done it, you need to take inventory of what you already own and document your findings. You will use this information for years to come as long as you keep it up to date.

  • What tools exist and what category should they be put in? (Database Monitoring, Network Monitoring, OS Monitoring, Desktop Monitoring, etc…)
  • How many licenses do we have and are the current?
  • What are they good at?
  • What are they not good at?
  • What would be classified as an APM tool?
  • If I already have an APM tool why is it not being used properly?
  • Put labels on your existing tools and understand what they do!

Down in a Hole – Alice In Chains

Now that you have the overall landscape of your monitoring ecosystem laid out you need to see if there are any gaping holes. AppDynamics is an APM company so we suggest you compare your existing tools to the Gartner definition of APM to see what you might be missing. Here it is paraphrased in my own words:

  • End User Experience Monitoring: Measuring the response time of your application all the way to the end user. It’s not good enough to just understand how fast your application runs within the confines of the data center(s).
  • Application Topology Mapping: Automatic detection and display of all components involved in the delivery of your application. You need to know what application components are in use at any given time, but especially when there is an issue impacting your users.
  • Business Transaction Profiling: Detecting and measuring the response time of all application component activity initiated by a single user request. This is not the same as measuring the response time of a web page!!!
  • Deep Application Diagnostics: Detecting and measuring the run time code execution within your application containers. If your current or prospective solution does not load into the application container you will NOT have this important capability.
  • Analytics: Intelligence applied to data which provides you with actionable information. This is not the same as reporting and analytics can (and should) be a key differentiator between competing solutions.

You can do this same type of gap analysis for other types of monitoring but you will need to figure out the main aspects of each type for yourself (sorry).

So What’cha Want – The Beastie Boys

So lets assume that you need an APM solution to fill that large void in your monitoring capabilities and to solve that pesky problem which has been lingering for the past 6 months. You know which problem I’m referring to, it’s the one that crops up for a few hours every other week, it wreaks havoc on your business, then it magically disappears after half of your IT staff is on a massive conference call trying to figure out what’s going on. It’s the problem that your boss has to answer for in the daily update calls and can never explain. It’s the problem that will eventually get someone fired or “re-organized” and if you can fix it you will be a hero or a rockstar (you get to choose which term you prefer since you fixed the problem).

In order to pick the right APM solution (picking the wrong one can turn that rockstar potential completely upside down) you need to develop a method for comparing different solutions. First you need to narrow down the crowded list of vendors to a group of 2 or 3 at most to do a Proof Of Concept (POC). This initial narrowing of the field is usually done through feature comparisons (vendor websites), references from people you know, phone calls with vendors, phone calls with analysts (Gartner, Forrester, etc…), crystal balls, voodoo rituals, and possibly animal sacrifice.

Use your favorite spreadsheet program (I always wonder how many people don’t actually use Excel at work) and create a matrix with all of the products you might be interested in (AppDynamics should be first on the list, wink, wink) and compare each product to the level of support they have for each of your requirements. Your requirements at this point should be pretty generic as the detailed requirements are best suited for the POC phase. Here are some requirements to get you started:

  • Automatic detection, naming, and monitoring of Business Transactions
  • Automatic discovery and deep instrumentation of application code
  • End user experience monitoring
  • Analytics based alerting
  • Automatic discovery and display of my application topology
  • Support for my application technologies
  • Support for my application architecture (cloud, monolithic, distributed, etc…)
  • Open-ness of vendor (did they skirt around my questions?, did they talk in circles?)
  • Many more requirements you should add…

One you have this matrix built it should be easy to narrow your choices (hopefully AppDynamics made the short list, wink, wink).

One piece of advice before moving along to the POC requirements section… Be very explicit and detailed with your questions as they apply to your environment. If you get a partial answer from a vendor or a response that is not quite what you are looking for make sure you dig deeper right away to figure out if they are blowing smoke or not. If they keep talking in circle you can be pretty sure they can’t really do what you are asking.

Little Lies – Fleetwood Mac

No, this is not the point where I tell you to start making stuff up so you can get your software purchased. This is the part where I warn you that vendors can get very “creative” with their marketing. Some vendors are worse about this than others but it can be very difficulty to wade through the enormous amounts of BS that are splattered across many vendors websites.

Remember this key principle when dealing with all vendors … Don’t trust anything a vendor tells you, make them show you in a live demo!!!

I’m not saying that vendors are evil liars, but they will give you their best answer to your questions and it might not align with the intent of your question. This has happened to me many times before and I learned the hard way that although a vendors answer was technically true, the answer was far from reality when compared with the intent of my question.

And now back to our regularly scheduled broadcast.

3 is the Magic Number – De La Soul

The POC demands it own set of much more detailed requirements. This is the time to really look under the covers at a product and see what it can do versus what you have already been told. Again, using your favorite spreadsheet tool create a new sheet for each tool in the POC. I prefer to use a weighting and grading scale to help differentiate between tools.

  • Weighting: Some requirements will be more important to you than others. I use a Low, Medium, and High scale with associated values of 1, 2, and 3. This is your multiplier for the grade you give each requirement
  • Grading: Some tools have better support for a requirement than others. I use a grading scale of None, Poor, Adequate, and Excellent with associated values of 0, 1, 2, and 3.

I also like to break up the POC requirements into categories. This helps keep me organized, ensures I don’t miss testing something on my list, and helps me write the analysis document at the end of the POC. Some examples of categories and requirements are shown below.

  • Authentication and Authorization
    • Supports Microsoft AD
    • Granular Role Based Permission System
  • User Interface
    • Web based user interface compatible with Internet Explorer 8+
    • Displays application topology without administrator or user configuration
  • Deep Application Diagnostics
    • Automatically discovers and instruments custom code
    • Automatically traces complete call stack when performance is abnormally poor
    • Automated intelligence to ensure instrumentation does not use excessive overhead

Your list should be much longer and more detailed. This requirements list is the basis of the POC and all follow up documentation so make sure it is thorough and not slanted towards a particular vendor. Think of it as your Christmas list when you were a kid, don’t be afraid to ask for things that might seem impossible but that could be really useful.

Another key lesson learned coming your way… Don’t let the vendor control the POC. You define the environment (Dev, Test, and even Prod if it can be done safely), you do the installation, you do the configuration, you do everything related to the POC. You are the one who has to use the tool after you buy it so be sure to personally do everything during the POC.

Be sure you provide the same playing field for each vendor so that your results are really comparable. Having a successful POC using a team of 5 vendor engineers versus another successful POC using just 1 engineer is comparing apples and oranges.

We Are the Champions – Queen

After you wrap up all of your POCs you should have enough data to pick a winner. Hopefully you built your spreadsheets so that they automatically add up the numbers related to all of the requirements. You will have a statistical winner based purely upon you spreadsheet data and that usually aligns with the overall feeling you have after the POC is complete but sometimes it might not. You may have a vendor/product that is statistically the winner but your gut is telling you to choose a different vendor. In this case you need to figure out exactly why you feel this way. You cannot justify vendor selection based upon a gut instinct with no explanation. Maybe one vendor was just a royal pain to deal with, maybe the solution worked well but was agonizingly painful to deploy. The key here is being able to express why your instinct is pushing you a certain direction and quantify the potential impact. Something similar to the following statement can be used as justification of your position:

“Even though Vendor X Product statistically scored highest in the evaluation there is one overriding factor that prohibits selecting them as the overall winner. Deployment and configuration of Vendor X Product is difficult and time consuming. Based upon the observed deployment and configuration time of 2 weeks for 1 application during the POC it would take approximately 19 years to configure monitoring for the 500 applications which are in scope. Deploying Vendor X Product does not make sense in our environment.”

The information you have been building throughout this entire process should be used to create an overall evaluation document. This document should have the following information at a minimum:

  • Description of problem
  • Description of proposed solution
  • Vendors/Products evaluated
  • Evaluation Criterial (Requirements)
  • Evaluation Results
  • Recommendation
  • Next steps

It can also be helpful to create a short (3–10 pages) presentation to accompany the evaluation document which you can use to brief management on your findings. This presentation should contain only the most important facts since you have all the detail anyone should ever want in the full evaluation document.

Another important document you can create is the business justification. The business justification cuts out all the technical details related to the product you want to purchase and gets right down to the economics of the matter. I am not going to dive deep into how to write a business justification but to help you get started you should make sure you get a ROI (Return On Investment) calculator from each vendor that participates in your POC. Vendors want to help you buy their solution and have a wealth of information available to help you build your business justification so just ask them for help if you need it. Just make sure that everything in the business justification is factually based and relevant to your business.

Lastly it is really helpful to have broad support for your initiative. Seek out people across your organization that will support your recommendation of product or who will validate the problem that you are trying to solve. If you have full support of an Application Owner whose business is being directly impacted this greatly increases your chances of success.

If you read this entire blog post please accept my gratitude! I know it was a long post but it’s a huge topic to cover. Hopefully you picked up some good information and will return for my next post about deploying the product you just worked so hard to get in the door. As always, your comments are welcome. I’d love to hear your tips for getting new solutions approved or any war stories from your vendor evaluations.

Read the original blog entry...

More Stories By Jyoti Bansal

In high-production environments where release cycles are measured in hours or minutes — not days or weeks — there's little room for mistakes and no room for confusion. Everyone has to understand what's happening, in real time, and have the means to do whatever is necessary to keep applications up and running optimally.

DevOps is a high-stakes world, but done well, it delivers the agility and performance to significantly impact business competitiveness.

@ThingsExpo Stories
Recently, WebRTC has a lot of eyes from market. The use cases of WebRTC are expanding - video chat, online education, online health care etc. Not only for human-to-human communication, but also IoT use cases such as machine to human use cases can be seen recently. One of the typical use-case is remote camera monitoring. With WebRTC, people can have interoperability and flexibility for deploying monitoring service. However, the benefit of WebRTC for IoT is not only its convenience and interopera...
There is only one world-class Cloud event on earth, and that is Cloud Expo – which returns to Silicon Valley for the 21st Cloud Expo at the Santa Clara Convention Center, October 31 - November 2, 2017. Every Global 2000 enterprise in the world is now integrating cloud computing in some form into its IT development and operations. Midsize and small businesses are also migrating to the cloud in increasing numbers. Companies are each developing their unique mix of cloud technologies and service...
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
SYS-CON Events announced today that App2Cloud will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct. 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. App2Cloud is an online Platform, specializing in migrating legacy applications to any Cloud Providers (AWS, Azure, Google Cloud).
IoT is at the core or many Digital Transformation initiatives with the goal of re-inventing a company's business model. We all agree that collecting relevant IoT data will result in massive amounts of data needing to be stored. However, with the rapid development of IoT devices and ongoing business model transformation, we are not able to predict the volume and growth of IoT data. And with the lack of IoT history, traditional methods of IT and infrastructure planning based on the past do not app...
To get the most out of their data, successful companies are not focusing on queries and data lakes, they are actively integrating analytics into their operations with a data-first application development approach. Real-time adjustments to improve revenues, reduce costs, or mitigate risk rely on applications that minimize latency on a variety of data sources. Jack Norris reviews best practices to show how companies develop, deploy, and dynamically update these applications and how this data-first...
Intelligent Automation is now one of the key business imperatives for CIOs and CISOs impacting all areas of business today. In his session at 21st Cloud Expo, Brian Boeggeman, VP Alliances & Partnerships at Ayehu, will talk about how business value is created and delivered through intelligent automation to today’s enterprises. The open ecosystem platform approach toward Intelligent Automation that Ayehu delivers to the market is core to enabling the creation of the self-driving enterprise.
Internet-of-Things discussions can end up either going down the consumer gadget rabbit hole or focused on the sort of data logging that industrial manufacturers have been doing forever. However, in fact, companies today are already using IoT data both to optimize their operational technology and to improve the experience of customer interactions in novel ways. In his session at @ThingsExpo, Gordon Haff, Red Hat Technology Evangelist, shared examples from a wide range of industries – including en...
Consumers increasingly expect their electronic "things" to be connected to smart phones, tablets and the Internet. When that thing happens to be a medical device, the risks and benefits of connectivity must be carefully weighed. Once the decision is made that connecting the device is beneficial, medical device manufacturers must design their products to maintain patient safety and prevent compromised personal health information in the face of cybersecurity threats. In his session at @ThingsExpo...
"We're a cybersecurity firm that specializes in engineering security solutions both at the software and hardware level. Security cannot be an after-the-fact afterthought, which is what it's become," stated Richard Blech, Chief Executive Officer at Secure Channels, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that Massive Networks will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Massive Networks mission is simple. To help your business operate seamlessly with fast, reliable, and secure internet and network solutions. Improve your customer's experience with outstanding connections to your cloud.
SYS-CON Events announced today that Grape Up will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct. 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Grape Up is a software company specializing in cloud native application development and professional services related to Cloud Foundry PaaS. With five expert teams that operate in various sectors of the market across the U.S. and Europe, Grape Up works with a variety of customers from emergi...
Detecting internal user threats in the Big Data eco-system is challenging and cumbersome. Many organizations monitor internal usage of the Big Data eco-system using a set of alerts. This is not a scalable process given the increase in the number of alerts with the accelerating growth in data volume and user base. Organizations are increasingly leveraging machine learning to monitor only those data elements that are sensitive and critical, autonomously establish monitoring policies, and to detect...
Everything run by electricity will eventually be connected to the Internet. Get ahead of the Internet of Things revolution and join Akvelon expert and IoT industry leader, Sergey Grebnov, in his session at @ThingsExpo, for an educational dive into the world of managing your home, workplace and all the devices they contain with the power of machine-based AI and intelligent Bot services for a completely streamlined experience.
Because IoT devices are deployed in mission-critical environments more than ever before, it’s increasingly imperative they be truly smart. IoT sensors simply stockpiling data isn’t useful. IoT must be artificially and naturally intelligent in order to provide more value In his session at @ThingsExpo, John Crupi, Vice President and Engineering System Architect at Greenwave Systems, will discuss how IoT artificial intelligence (AI) can be carried out via edge analytics and machine learning techn...
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
With tough new regulations coming to Europe on data privacy in May 2018, Calligo will explain why in reality the effect is global and transforms how you consider critical data. EU GDPR fundamentally rewrites the rules for cloud, Big Data and IoT. In his session at 21st Cloud Expo, Adam Ryan, Vice President and General Manager EMEA at Calligo, will examine the regulations and provide insight on how it affects technology, challenges the established rules and will usher in new levels of diligence a...
An increasing number of companies are creating products that combine data with analytical capabilities. Running interactive queries on Big Data requires complex architectures to store and query data effectively, typically involving data streams, an choosing efficient file format/database and multiple independent systems that are tied together through custom-engineered pipelines. In his session at @BigDataExpo at @ThingsExpo, Tomer Levi, a senior software engineer at Intel’s Advanced Analytics ...
In the enterprise today, connected IoT devices are everywhere – both inside and outside corporate environments. The need to identify, manage, control and secure a quickly growing web of connections and outside devices is making the already challenging task of security even more important, and onerous. In his session at @ThingsExpo, Rich Boyer, CISO and Chief Architect for Security at NTT i3, discussed new ways of thinking and the approaches needed to address the emerging challenges of security i...
SYS-CON Events announced today that Dasher Technologies will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 - Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Dasher Technologies, Inc. ® is a premier IT solution provider that delivers expert technical resources along with trusted account executives to architect and deliver complete IT solutions and services to help our clients execute their goals, plans and objectives. Since 1999, we'v...