Everyone’s in the cloud,” they said, “and if you’re not there, your business will die.

That was, when, 2010?

As it turned out, there are too many moving parts, too many applications weren’t ready, bandwidth was too expensive and too unreliable, and, oh, storing sensitive data elsewhere? Move along Sir, nothing to see here.

It took us ten years, which is like a century in tech-years (2010: iPhone 4, anyone?) to get to the current state.


2020 is approaching, and everyone is in the cloud.


Just in a different way than expected. And not dead, anyway.

While there are loads of businesses operating exclusively in the cloud, the majority are following the hybrid approach.

Hybrid is easier for us folks here in EMEA, where each country still follows their laws, while at the same time sitting under the GDPR umbrella.
But let’s not overcomplicate things. Hybrid is simply the best of both worlds.


There are loads of cloud providers, and according to a survey we ran recently with our THWACK® community, Microsoft® Azure® is the most popular one in our customer base. Around 53% of you are using Azure, one way or the other.

We added monitoring support for Azure into the Orion® Platform with Server & Application Monitor (SAM) version 6.5 in late 2017, and since, it’s been possible to monitor servers deployed in the cloud and to retrieve other supporting stats, for example, storage.

BTW, the iPhone 8 was the latest tech back then—to put things into perspective.

Earlier in the year, we added Azure SQL support to Database Performance Analyzer (DPA). DPA was our first product available in the Azure Marketplace—it’s been out there for a while.


The question for deploying the Orion Platform in Azure came up frequently.

We heard some interesting stories—true pioneers who run the Orion Platform in the cloud, completely unsupported and “at your own risk,” but hey, it worked.

So, the need was there, and we heard you: Azure deployment became officially supported in 2018, and in early 2019 we added some bits and pieces and created the documentation.

But we identified further room for improvement, partnered with Microsoft, and a few weeks after the iPhone 11 Pro release…I mean, since our latest 2019.4 release, the Orion Platform became available in the Azure Marketplace, and Azure SQL Managed Instances were officially supported in the Orion database.

The next step is obvious. It doesn’t matter where your workloads are located anymore, so why would you be concerned with the location of your monitoring system?


Let’s have a look at the present…


The speed and usability of your first Orion Platform deployment is improved through the marketplace, as the applications are pre-packaged and so much more than the plain executable you would run on a local server, or even an ISO file.

They’re tested and validated instances with all the necessary dependencies, offering options for customization and flexibility.

Traditionally you’d need to request host resources, clone a VM, right-size it a couple of times, make sure the OS is ready, install dependencies, and install the Orion Platform.

While none of it is rocket science, it’ll take some time, and there’s a risk in something going wrong. You know it’s going to happen.

But not anymore!


And in the future…


If one thing’s consistent, it’s change. And the change could very well mean you’re migrating the Orion Platform into Azure.
Previously, it was a semi-complex task and probably started with using our own Virtualization Manager and its sprawl-feature to make sure all machines involved are right-sized before migration to prevent costly surprises.

The next step is dealing with the infrastructure, the OS, double-clicking the executable again…I wouldn’t call it a grind, but it’s a grind.

And now?
A few clicks in the Azure Marketplace and the whole thing is deployed, and really all that’s left to do is to take care of moving the database. Alright, I agree, that’s a bit of a simplified view, but you know where I’m going with this. It’s easier. Big time.


Keyword database.
It’s my least favourite topic and I’m usually glad when Tom is around to discuss it, but from checking the current requirements, SQL 2014 is the minimum, but please consider some Orion Platform modules require 2016SP1 because of the column store feature.

Running anything older than this version doesn’t make a lot of sense.
Getting new licenses and deploying a new SQL version while at the same time paying attention to possible dependencies could be a roadblock preventing you from upgrading Orion modules, and even this is improved with Azure.


Also, it’s easier to test a new version, try a new feature, or even another module.
Why not run a permanent lab environment in the cloud? Just a thought!

Do you want to know more? Sure you do!


Here’s the next step for you, while I’m on a date with my next cup of coffee.

Omar Rafik, SolarWinds Senior Manager, Federal Sales Engineering


Here’s an interesting article by my colleague Brandon Shopp where he discusses the Army’s new training curriculum and the impacts on network and bandwidth.


The U.S. Army is undergoing a major technology shift affecting how soldiers prepare for battle. Core to the Army’s modernization effort is the implementation of a Synthetic Training Environment (STE) combining many different performance-demanding components, including virtual reality and training simulation software.


The STE’s One World Terrain (OWT) concept is comprised of five different phases from the initial point of data collection to final application. During phase four, data is delivered to wherever soldiers are training. Raw data is used to automatically replicate digital 3-D terrains, so soldiers can experience potential combat situations in virtual reality through the Army’s OWT platform before setting foot on a battlefield.


Making One World Terrain work


For the STE to work as expected, the Army’s IT team should consider implementing an advanced form of network monitoring focused specifically on bandwidth optimization. The Army’s objective with OWT is to provide soldiers with as accurate a representation of actual terrain as possible, right down to extremely lifelike road structures and vegetation. Transmitting so much information can create network performance issues and bottlenecks. IT managers must be able to continually track performance and usage patterns to ensure their networks can handle the traffic.


With this practice administrators may discover what areas can be optimized to accommodate the rising bandwidth needs presented by the OWT. For example, their monitoring may uncover other applications, outside of those used by the STE, unnecessarily using large amounts of bandwidth. They can shut those down, limit access, or perform other tasks to increase their bandwidth allocation, relieve congestion, and improve network performance, not just regarding STE resources but across the board.


Delivering a Consistent User Experience


But potential hidden components found in every complex IT infrastructure could play havoc with the network’s ability to deliver the desired user experience. There might be multiple tactical or common ally networks, ISPs, agencies, and more, all competing for resources and putting strain on the system. Byzantine application stacks can include solutions from multiple vendors, not all of which may play nice with each other. Each of these can create their own problems, from server errors to application failures, and can directly affect the information provided to soldiers in training.


To ensure a consistent and reliable experience, administrators should take a deep dive into their infrastructure. Monitoring database performance is a good starting point because it allows teams to identify and resolve issues causing suboptimal performance. Server monitoring is also ideal, especially if it can monitor servers across multiple environments, including private, public, and hybrid clouds.


These practices should be complemented with detailed application monitoring to provide a clear view of all the applications within the Army’s stack. Stacks tend to be complicated and sprawling, and when one application fails, the others are affected. Gaining unfettered insight into the performance of the entire stack can ward off problems that may adversely affect the training environment.


Through Training and Beyond


These recommendations can help well beyond the STE. The Army is clearly a long way from the days of using bugle calls, flags, and radios for communication and intelligence. Troops now have access to a wealth of information to help them be more intelligent, efficient, and tactical, but they need reliable network operations to receive the information. As such, advanced network monitoring can help them prepare for what awaits them in battle—but it can also support them once they get there.


Find the full article on Government Computer News.


The SolarWinds trademarks, service marks, and logos are the exclusive property of SolarWinds Worldwide, LLC or its affiliates. All other trademarks are the property of their respective owners.

For some, artificial intelligence (AI) can be a scary technology. There are so many articles on the web about how AI will end up replacing X% of IT jobs by Y year. There’s no reason to be afraid of AI or machine learning. If anything, most IT jobs will benefit from AI/machine learning. The tech world is always changing, and AI is becoming a big driver of change. Lots of people interact with or use AI without even realizing it. Sometimes I marvel at the video games my kids are playing and think back to when I played Super Mario Brothers on the original Nintendo Entertainment System (which I still have!). I ask myself, “What kind of video games will my kids be playing?” The same applies to the tech world now. What kinds of things will AI and machine learning have changed in 10 years? What about 20 years?


Let’s look at how AI is changing the tech world and how it will continue to benefit us in decades to come.


Making IoT Great

The idea of the Internet of Things focuses on all the things in our world capable of connecting to the internet. Our cars, phones, homes, dishwashers, washing machines, watches, and yes, even our refrigerators. The internet-connected devices we use need to do what we want, when we want, and do it all effectively. AI and machine learning help IoT devices perform their services effectively and efficiently. Our devices collect and analyze so much data, and they rely more and more on AI and machine learning to sift through and analyze all the data to make our interactions better.


Customer Service

I don’t know a single human being who gets a warm and fuzzy feeling when thinking about customer service. No one wants to deal with customer service, especially when it comes to getting help over the phone. What’s our experience today? We call customer service and we likely get a robotic voice (IVR) routing us to the next most helpful robot. I find myself yelling “representative” over and over until I get a human… really, it works!


AI is improving our experience with IVR systems by making them easier to interact with and get help from. IVR systems also use AI to analyze input from callers to better route their calls to the right queue based on common trending issues. AI also helps ensure 24/7 customer service, which can be helpful with off-hours issues. You don’t have to feed AI-enhanced IVR systems junk food and caffeine to get through the night!


Getting Info to the User Faster

Have you noticed the recommendations you get on YouTube? What about the “because you watched…” on Netflix? AI and machine learning are changing the way we get info. Analytical engines pour through user data and match users with the info they most want, and quickly. On the slightly darker side of this technology, phones and smart speakers have become hot mics. If you’re talking about how hungry you are, your phone or speaker hears you, then sends you an email or pops up an ad on your navigation system to the nearest restaurant. Is that good? I’m not sold on it yet—it feels a little invasive. Like it or not, the way we get our data is changing because of AI.


Embrace It or Pump the Brakes?

For some, yeah you know who I’m talking about (off-gridders), AI is an invasive and non-voluntary technology. I wouldn’t be surprised if you had to interact with AI at least once in your day. We can’t get around some of those moments. What about interacting with AI on purpose? Do you automate your home? Surf the web from your fridge door? Is AI really on track with helping us as humans function easier? It’s still up for debate, and I’d love to hear your response in the comments.

Not too long ago, a copy of Randall Munroe’s “Thing Explainer” made its way around the SolarWinds office—passing from engineering to marketing to development to the Head Geeks, and even to management.


Amid chuckles of appreciation, we recognized Munroe had struck upon a deeper truth: as IT practitioners, we’re often asked to describe complex technical ideas or solutions. However, it’s often for an audience requiring a simplified explanation. These may be people who consider themselves “non-technical,” but just as easily, it could be for folks deeply technical in a different IT discipline. From both groups (and people somewhere in-between) comes the request to “explain it to me like I’m five years old” (a phrase shortened to “Explain Like I’m Five,” or ELI5, in forums across the internet).


There, amid Munroe’s mock blueprints and stick figures, were explanations of complex concepts in hyper-simplified language achieving the impossible alchemy of being amusing, engaging, and accurate.


We were inspired. And so, for the December Writing Challenge 2019, we hope to do for IT what Randall Munroe did for rockets, microwaves, and cell phones: explain what they are, what they do, and how they work in terms anyone can understand, and in a way that may even inspire a laugh or two.


At the same time, we hope to demonstrate a simple idea best explained by a man who understood complicated things:

“If you can’t explain it simply, you don’t understand it well enough.” – Albert Einstein


Throughout December, one writer—chosen from among the SolarWinds staff and THWACK MVPs—will be the lead writer each day. You—the THWACK community—are invited to contribute your own thoughts each day, both on the lead post and the word itself. In return, you’ll receive friendship, camaraderie, and THWACK points. 200 to be precise, for each day you comment on a post.*


You’ll find each day’s post on the December Writing Challenge 2019 forum. Take a moment now to visit it and click "Follow" so that you don't miss a single post. As in past years, I’ll be writing a summary of the week and posting it over on the Geek Speak forum.


In the spirit of ELI5, your comments (and indeed, those of the lead writers as well) can be in the form of prose, poetry, or even pictures. Whatever you feel addresses the word of the day and represents a way to explain a complex idea simply and clearly.


To help you get your creative juices flowing, here’s the word list in advance.


Everyone here on THWACK is looking forward to reading your thoughts!

  1. Monitoring
  2. Latency
  3. Metrics
  4. NetFlow
  5. Logging
  6. Observability
  7. Troubleshoot
  8. Virtualization
  9. Cloud Migration
  10. Container
  11. Orchestration
  12. Microservices
  13. Alert
  14. Event Correlation
  15. Application Programming Interface (API)
  16. SNMP
  17. Syslog
  18. Parent-Child
  19. Tracing
  20. Information Security
  21. Routing
  22. Ping
  23. IOPS
  24. Virtual Private Network (VPN)
  25. Telemetry
  26. Key Performance Indicator (KPI)
  27. Root Cause Analysis
  28. Software Defined Network (SDN)
  29. Anomaly detection
  30. AIOps
  31. Ransomware


* We’re all reasonable people here. When I say “a comment,” it needs to be meaningful. Something more than “Nice” or “FIRST!” or “Gimme my points.” But I’m sure you all knew that already.

It’s time to stop seeing IT as just a black hole to throw money towards and instead, give back to the business. Often, the lack of proper measurement and standardization are the problem, so how do we address this?


We’ve all tried to leverage more funds from the organization to undertake a project. Whether replacing aging hardware or out-of-support software, without the business seeing the value the IT department brings, it’s hard to obtain the financing required. That’s because the board typically sees IT as a cost center.


A cost center is defined as a department within a business not directly adding profit, but still requiring money to run. A profit center is the inverse, whereby its operation directly adds to overall company profitability.


But how can you get there? You need to understand everything the IT department does is a process, and thereby can be measured and thus improved upon. Combining people and technology, along with other assets, and passing them through a process to deliver customer success or business results is a strategic outcome you want to enhance. If you can improve the technology and processes, then you’ll increase the strategic outcomes.


By measuring and benchmarking these processes, you can illustrate the increases made due to upgrades. Maybe a web server can now support 20% more traffic, or its loading latency has been reduced by four seconds, which over a three-month period has led to an increase of 15% web traffic and an 8% rise in sales. While I’ve just pulled these figures from the air, the finance department evaluates spending and can now see a tangible return on the released funds. The web server project increased web traffic, sales, and therefore profits. When you approach the next project you want to undertake, you don’t have to just say “it’s faster/newer/bigger than our current system” (although you should add to your discussion the faster, newer, bigger piece of hardware or software will provide an improvement to the overall system). However, without data, you have no proof, and the proof is in the pudding, as they say. Nothing will make a CFO happier than seeing a project with milestones and KPIs (well, maybe three quarters exceeding predictions). So, how do we measure and report all these statistics?


If you think of deployments in terms of weeks or months, you’re trying to deploy something monolithic yet composed of many moving and complex parts. Try to break this down. Think of it like a website. You can update one page without having to do the whole site. Then you start to think “instead of the whole page, what about changing a .jpg, or a background?” Before long, you’ve started to decouple the application at strategic points, allowing for independent improvement. At this stage, I’d reference the Cloud Native Computing Foundation Trail Map as a great way to see where to go. Their whole ethos on empowering organizations running modern scalable applications can help you with transformation.


But we’re currently looking at the measurement aspect of any application deployment. I’m not just talking about hardware or network monitoring, but a method of obtaining baselines and peak loads, and of being able to predict when a system will reach capacity and how to react to it.


Instead of being a reactive IT department, you suddenly become more proactive. Being able to assign your ticketing system to your developers allows them to react faster to any errors from a code change and quickly fix the issue or revert to an earlier deployment, thereby failing quickly and optimizing performance.


I suggest if you’re in charge of an application or applications, or support one on a daily basis, start to measure and record anywhere you can on the full stack. Understand what normal looks like, or how it’s different from “rush hour,” so you can say with more certainty it’s not the network. Maybe it’s the application, or it’s the DNS (it’s always DNS) leading to delays, lost revenue, or worse, complete outages. Prove to the naysayers in your company you have the correct details and that 73.6% of the statistics you present aren’t made up.

Continuing from my previous blog post, Meh, CapEx, I’m going to take a cynical look at how and why Microsoft has killed its perpetual licensing model. Now don’t get me wrong, it’s not just Microsoft – other vendors have done the same. I think a lot of folks in IT can say they use at least one Microsoft product, so it’s easily relatable.



Let’s start with the poster child for SaaS done right: Office 365. Office 365 isn’t merely a set of desktop applications like Word and Excel with such cool features as Clippy anymore.


No, it’s a full suite of services such as email, content collaboration, instant messaging, unified communications, and many more, but you already knew that, right? With a base of 180 million active users as of Q3 2019¹ and counting, it’d be silly for Microsoft not to invest their time and effort into developing the O365 platform. Traditional on-premises apps, though, are lagging in feature parity or in some cases changed in a way that to me, at least, seems like a blatant move to push people towards Office 365. Let’s look at the minimum hardware requirements for Exchange 2019 for example: 128GB memory required for the mailbox server role². ONE HUNDRED AND TWENTY-EIGHT! That’s a 16 times increase over Exchange 2016³. What’s that about then?


To me, it seems like a move to guide people down the path of O365. People without the infrastructure to deploy Exchange 2019 likely have a small enough mail footprint to easily move to O365.


Like I said in my Meh, CapEx blog post, it’s the extras bundled in with the OpEx model make it even more attractive. Microsoft Teams is one such example of a great tool that comes with O365 and O365 only. Its predecessors, Skype for Business and Lync on-premises, are dead.


  Now, what about Microsoft Azure? Check out this snippet from the updated licensing terms as of October 1, 2019:

Beginning October 1, 2019, on-premises licenses purchased without Software Assurance and mobility rights cannot be deployed with dedicated hosted cloud services offered by the following public cloud providers: Microsoft, Alibaba, Amazon (including VMware Cloud on AWS), and Google.

So basically, no more perpetual license on one of the big public cloud providers for you, Mr./Mrs. Customer.


Does this affect you? I’d love to know.


I saw some stats from one of the largest Microsoft distributors in the U.K., 49% of all deployed workloads in Azure that are part of a CSP subscription they’ve sold are virtual machines. I’d be astonished if this license change doesn’t affect a few of those customers.


Wrap It Up

In my cynical view, Microsoft is leading you down a path where subscription licensing is more favorable. You only get the cool stuff with a subscription license, while traditional on-premises services are being made to look less favorable one way or another. And guess what—they were usually licensed with a perpetual license.


It’s not all doom and gloom though. Moving to services like O365 also removes the headache of having to manage services like Exchange and SharePoint. But you must keep on paying, every month, to continue to use those services.



¹ Microsoft third quarter earnings call transcript, page 3 https://view.officeapps.live.com/op/view.aspx?src=https://c.s-microsoft.com/en-us/CMSFiles/TranscriptFY19Q3.docx?version=0e85483a-1f8a-5292-de43-397ba1bfa48b


² Exchange 2019 system requirements https://docs.microsoft.com/en-us/exchange/plan-and-deploy/system-requirements?view=exchserver-2019


³ Exchange 2016 system requirements https://docs.microsoft.com/en-us/exchange/plan-and-deploy/system-requirements?view=exchserver-2016


⁴ Source, Microsoft licensing terms for dedicated cloud https://www.microsoft.com/en-us/licensing/news/updated-licensing-rights-for-dedicated-cloud

I am back in Orlando this week for Live 360, where I get to meet up with 1,100 of my close personal data friends. If you're attending this event, please find me--I'm the tall guy who smells like bacon.


As always, here are some links I hope you find interesting. Enjoy!


Google will offer checking accounts, says it won’t sell the data

Because Google has proved itself trustworthy over the years, right?


Google Denies It’s Using Private Health Data for AI Research

As I was just saying...


Automation could replace up to 800 million jobs by 2035

Yes, the people holding those jobs will transition to different roles. It's not as if we'll have 800 million people unemployed.


Venice floods: Climate change behind highest tide in 50 years, says mayor

I honestly wouldn't know if Venice was flooded or not.


Twitter to ban all political advertising, raising pressure on Facebook

Your move, Zuck.


California man runs for governor to test Facebook rules on lying

Zuckerberg is doubling down with his stubbornness on political ads. That's probably because Facebook revenue comes from such ads, so ending them would kill his bottom line.


The Apple Card Is Sexist. Blaming the Algorithm Is Proof.

Apple, and their partners, continue to lower the bar for software.


Either your oyster bar has a trough or you're doing it wrong. Lee & Rick's in Orlando is a must if you are in the area.


Omar Rafik, SolarWinds Senior Manager, Federal Sales Engineering


Here’s an interesting article by Mark Hensch about when my colleague Arthur Bradway spoke at a conference about gamifying cybersecurity to improve results.


Can gaming significantly improve how governments protect their cybersecurity by making their employees more careful about how they use their IT?


According to one cybersecurity expert, the answer might be yes. Arthur Bradway says turning security training into a game can help public servants remember tips for keeping their agency’s data safe.


“A big topic lately is gamifying the security training,” Bradway said during GovLoop’s virtual summit. “We all like games. We all like to win.”


Bradway is a senior government sales engineer at SolarWinds, a software provider specializing in network, systems, and IT management.


According to Bradway, many agencies use dull training lectures, presentations, and videos that don’t help their employees retain cybersecurity knowledge.


“A lot of these methods aren’t engaging to the end user,” he said. “You spend time to get your users there and they don’t remember anything. By making it more engaging, they’ll retain more of the information.”


Bradway said creating games can also help agencies establish, teach, and enforce IT controls for their workforces.


IT controls consist of the procedures and policies to help ensure technologies are being used for their intended purposes in a reasonable manner.


Examples of some general controls used for essential IT processes include risk and change management, security, and disaster recovery.


When it comes to IT controls, government employees are often unaware of what their agencies expect from them in terms of cybersecurity.


“End users are our weakest links in all of this,” he said. “The majority of them don’t know anything about security. They’re used to being constantly connected anywhere they want on their devices. They assume they’ll be able to do the same thing at the office.”


Unfortunately, governments can’t take cybersecurity concerns lightly because of the sensitive data they often handle.


Governments that fail to protect their data can lose the trust of their citizens, suffer financial damage, and even endanger national security.


Bradway said, however, gaming can help prevent cybersecurity incidents by teaching public servants about the topic in an entertaining way.


For example, he continued, gaming can educate people about the different cyberthreats currently menacing agencies.


Bradway suggested one game where players assume the role of such cyberthreats as hostile foreign governments to learn how they act.


“When people are playing the role of the bad guy, they realize, ‘Wait, there’s more than one type of bad guy in the world?’” he said. “They realize more is going on and they need to start paying attention to it.”


Gamifying security training could resonate with public servants—especially younger ones—who are used to playing games on their mobile devices.


“Everyone is used to doing something on their phones and getting some little reward,” he said. “We know the end users are the problem. A lot of this highlights the trainings, policies, and procedures in place.”


Find the full article on GovLoop.


The SolarWinds trademarks, service marks, and logos are the exclusive property of SolarWinds Worldwide, LLC or its affiliates. All other trademarks are the property of their respective owners.

The reports of my death are greatly exaggerated - The On Premises Data Center Mark Twain


As I mentioned in my previous post, the hype cycle has the on-premises data center dead to rights. Public cloud is the future and there’s no turning back. When you step back and investigate the private cloud space, you’d be surprised.


When the order comes down to move to the public cloud, sometimes you discover the application or workload isn’t suitable for migration. How many times have you been in an environment where an important line-of-business application was written by an employee who left 15 years ago and has no documentation. Is the application even running on an operating system one of the public cloud providers supports? While proper planning and refactoring can provide a path to move your application to the cloud, occasionally you’ll run into unavoidable physical limitations. In manufacturing environments, often you’ll find the need for a custom interface for a machine or a USB license dongle for an application to function.


Sometimes applications can’t move to the cloud despite planning. But not everyone takes the time to plan their cloud migrations. Failure to plan leads to many issues with cloud adoption (which I’ll discuss in my next post). What about your connectivity to the internet and cloud? Whether it’s uptime or bandwidth, the network can prove to be a cloud roadblock. When these migrations fail, where does the workload go? Back to the on-premises data center!


Looking at purchasing new data center hardware or moving to the cloud often includes decision-makers outside of the IT department. While we will leave the deep fiscal details for the financial experts, Chief Financial Officers and business leaders often weigh the benefits of the Operating Expenditure versus Capital Expenditures, commonly referred to as OpEx vs. CapEx. While the ability to quickly scale your cloud resources up and down based on demand might be a blessing to an application administrator, the variations on cost accompanying it can prove to be difficult for the accounting department. The ability to make a one-time purchase every three to five years and amortize the cost over those years is a tried-and-true financial tactic.


Speaking of finances, the fiscal performance of hardware vendors must certainly be a bellwether of cloud adoption. Sales of data center hardware has to be falling with the ever-growing adoption of public cloud, right? Wrong. As has been announced over the past year, vendors such as Dell Technologies, Pure Storage, and Lenovo report record earnings and growth. In May 2019, Dell Technologies announced 2% growth year over year. May also brought a revenue announcement from Lenovo of 12.5% growth year over year. August of 2019 saw Pure Storage announce a whopping 28% year over year revenue growth. These companies are just a small example. Clearly physical data center hardware is still in high demand.


Looking at many factors, it’s easy to say the on-premises data center is far from dead. Continuing the “Battle of the Clouds” series, we’ll dive into why “just put it in the cloud” isn’t always the best solution. 



Few relationships sail along without ever having to cross stormy seas. Even the best marriages need to batten down the hatches occasionally. A combination of genuine involvement, good communication, and various tools used by all involved is the best way to extend the time between encountering tempests. Business and professional relationships are no different.


Calming the Waters

The relationship between an IT department and the business it supports is one of provider and customer. A real relationship thrives when all participants are genuinely involved and prospers with good communication. Set up adequately and then maintained, IT Service Management (ITSM) is designed to create an environment where the relationship flourishes.


What Is ITSM?

The set of activities involved to plan, design, create, deliver, operate, and manage IT services for the business.


What Does This Mean?

Most people are familiar with the paradigm where IT systems management focuses on managing specific technology silos such as network, storage, and compute. Success is measured in uptime, availability, utilization, and how often the email inbox doesn’t get flooded. IT service management focuses on the processes to meet employee needs and business requirements. Success is measured by employee satisfaction and quality of experience. The processes for ITSM maintain a keen focus and a strong motivation for continuous improvement. An email or chat notification is an opportunity to jump into action and enhance the relationship.


Benefits of a Healthy Relationship

Great things can happen when IT is in a healthy relationship with the broader organization and employees. ITSM can deliver benefits for IT and the business.


Benefits Realized by IT:

  • A better understanding of business requirements
  • Well-defined processes
  • Proactive problem resolution
  • Improved tracking of repeat problems
  • Stats to measure and improve performance
  • Defined roles and responsibilities
  • Happier employees


Benefits Realized by the Business:

  • Increased IT service availability translates to increased employee and business productivity
  • Increased value and cost-efficiency
  • Reduced risk as the business complies with industry regulations
  • Adoption of new IT technology at a quicker pace
  • Increased competitive advantage


One Size Does Not Fit All

An IT department can be set up in many ways, and a business can consume the services offered by IT in many ways. Because of these variables, there are myriad types of relationships between the business and IT. It only makes sense that there are multiple ITSM frameworks to choose from.


Some Common Frameworks:

  • Business Process Framework (eTOM)
  • Control Objectives for information and Related Technologies (COBIT)
  • FitSM
  • ISO/IEC 20000
  • ITIL
  • Lean
  • Six Sigma
  • The Open Group Architecture Framework (TOGAF)


Supporting the business and helping it achieve its goals is true north for any IT department. An ITSM framework should, therefore, be chosen with the business in mind and for business reasons. It’s possible to build your own by adopting the most relevant parts from some or all of the frameworks listed above. It’s also possible to incorporate homegrown processes into one of the well-known frameworks. This mixing and matching allow complete customization of the way IT offers its services.


What’s in a Framework?

The success or failure of any system can usually be traced back to people, processes, and tools; in order of importance. While the IT department supplies the people and the tools, the ITSM framework delivers on the processes. All processes are designed to shift the habits of IT from being operationally focused on systems to being employee-focused and service-oriented.There are conventional processes contained in each of the different frameworks. The broad categories for these processes are:


  • Service Desk
  • Incident Management
  • Problem Management
  • Change Management
  • Asset Management
  • Knowledge Management


The Core of ITSM

A core benefit of ITSM is how it can positively change the relationship between the business and IT. When IT takes on a more employee-focused and service-oriented approach to delivering its services, IT can transform from being viewed merely as a cost center to a trusted business partner. There will always be a need for IT to be tactical in certain situations. By adopting and being genuinely involved in ITSM processes, IT can spend more time being strategic. IT can be a business enabler.

We’ve all heard it before.


“The cloud is the future!”

“We need to move to the cloud!”

“The on-premises data center is dead.”


If you believe the analysts and marketing departments, public cloud is the greatest thing to happen to the data center since virtualization. But, is it true? Could public cloud be the savior of the IT department? While many features to the public cloud make it an attractive infrastructure replacement, failure to adequately plan for its use can prove to be a costly mistake.


Moving past the marketing, the cloud is simply “someone else’s computer.” Yes, it’s more complicated than that, but when you boil it down to the basics, it’s a data center maintained by a third-party with proprietary software on top to provide an easy-to-use dashboard for provisioning and monitoring. When you move to the cloud, you’re still running an application on a server. Many of the same problems you have with your application running on-premises can persist in the cloud.


In a public cloud environment, the added complexity of multi-tenancy on the underlying resources can complicate things. Now you have to think about regulatory compliance? And after all, public cloud is still a data center subject to human error. This has been made evident over and over, famously by the Amazon Web Services S3 outage of February 2017.* The wide adoption of public clouds such as AWS and Microsoft Azure has also opened the door to more instances of shadow IT. Rogue devs, admins, and end users who either don’t have the patience to wait or have been denied resources opening cloud accounts with their own credit cards and putting corporate data at risk. And, we have yet to even take into consideration the consumption-based billing model.


Even with the above listed “issues” (I put quotes around issues as some of the problems can be encountered in the private cloud or worked around), public cloud can be an awesome tool in the IT administrator’s toolbox. Properly architected cloud-based applications can alleviate performance issues and can be developed with robust redundancies to avoid downtime. The ability to quickly scale compute up and down based on demand provides the business amazing agility not before seen in the standard data center procurement cycle. And, the growing world of SaaS products provides an easy gateway to enter the cloud (yes, I’m going to take the stance that as-a-Service qualifies as cloud). The introduction of cloud technologies has also opened a world of new application deployment models such as microservices and serverless computing. These amazing ways of looking at infrastructure weren’t possible until recently.


Is there hype around public cloud? For sure! Is some of it warranted? Absolutely! Is it the be-all and end-all technology of the future? Not so fast. In the upcoming series of posts I’m calling “Battle of the Clouds,” we’ll look at public cloud versus private cloud, going past the hype to dive into the state of on-premises data centers, what it takes for a successful cloud implementation, and workload planning around both solutions. I look forward to hearing your opinions on this topic as well!


*Summary of the Amazon S3 Service Disruption in the Northern Virginia (US-EAST-1) Region

Omar Rafik, SolarWinds Senior Manager, Federal Sales Engineering


Here’s an interesting article by my colleague Brandon Shopp with ideas for protecting IT assets in battlefield situations.


“Dominance” and “protection” sum up the Defense Department’s goals as U.S. armed forces begin to modernize their networks and communications systems. DOD is investing significant resources in providing troops with highly advanced technology so they can effectively communicate with each other and allies in even the harshest environments.


Efforts like the Army’s ESB-E tactical network initiative, for example, represent an attempt to keep warfighters constantly connected through a unified communications network. These solutions will be built off more scalable, adaptable, and powerful platforms than those provided by older legacy systems.


Programs like ESB-E are being designed to provide wide-scale communications in hostile territory. It will be incumbent upon troops in the field to monitor, manage, and secure the network to fulfill the “protection” part of DOD’s two-fisted battlefield domination strategy.


Moving forward to take this technological hill, DOD should keep these three considerations in mind.


  1. 1. The Attack Surface Will Increase Exponentially


Over the years, the battlefield has become increasingly kinetic and dependent upon interconnected devices and even artificial intelligence. The Army Research Laboratory calls this the internet of battlefield things—a warzone with different points of contact ultimately resulting in everything and everyone being more connected and, thus, intelligent.


The Pentagon is looking to take the concept as far as possible to give warfighters a tactical and strategic edge. For example, the Army wants to network soldiers and their weapons systems, and the Navy plans to link its platforms across hundreds of ships.


Opening these communication channels will significantly increase the potential attack surface. The more connection points, the greater the threat of exposure. Securing a communications system of such complexity will prove to be a far more daunting challenge than what’s involved in monitoring and managing a traditional IT network. Armed forces must be prepared to monitor, maintain, and secure the entire communications system.


  1. 2. Everyone Must Have Systems Expertise


The line between soldiers and system administrators has blurred as technology has advanced into the battlefield. As communications systems expand, all service members must be able to identify problems to ensure both unimpeded and uninterrupted communications and the security of the information being exchanged.


All troops must be bought into the concept of protecting the network and its communications components and be highly skilled in managing and maintaining these technologies. This is particularly important as communications solutions evolve.


Soldiers will need to quickly secure communications tools if they’re compromised, just as they would any other piece of equipment harboring sensitive information or access points. And they will require clear visibility into the entirety of the network to be able to quickly pinpoint any anomalies.


  1. 3. Staff Must Increase Commensurate to the Size of the Task


The armed forces must bulk up on staff to support these expansive modern communications systems. Fortunately, the military has a wealth of individuals with network and systems administration experience. Unfortunately, they lack in other critical areas.


Security specialists remain in high demand, but the cybersecurity workforce gap is real, even in the military. The White House’s National Cyber Strategy offers some good recommendations, including reskilling workers from other disciplines and identifying and fostering new talent. The actions highlighted in the plan coalesce with DOD’s need to fortify and strengthen its cybersecurity workforce as it turns its focus toward relentlessly winning the battlefield communications war.


Whoever wins this war will truly establish dominance over air, land, sea, and cyberspace. Victory lies in educating and finding the right personnel to protect information across what will undoubtedly be a wider and more attractive target for America’s adversaries.


Find the full article on Government Computer News.


  The SolarWinds trademarks, service marks, and logos are the exclusive property of SolarWinds Worldwide, LLC or its affiliates. All other trademarks are the property of their respective owners.

We talk a lot about professional development these days—how to develop your speaking or writing skills to make you a better practitioner. But what if you’ve reached a certain point in your career and you feel it’s time for a change?


Do you stay a technologist, or do you make the leap to management? It’s a conversation I’ve had with many people over the years and a question I’ve wrestled with myself. Several years ago, I made the jump. At the time, it was the wrong choice. In the following years, I spoke with many leaders and gathered many thoughts on what makes for a successful and satisfied IT leader. Below are the top five considerations I wish someone had shared with me prior to taking the management path.


How in Love With Technology Are You?

We could debate the best course, but often the person promoted to management is the one who knows the most about their technology. Ironically, once in management, you aren’t as hands-on as you once were. If all you want to do is geek out on tech, I’d urge you to be cautious about moving into management. Talk with your boss and other leaders to understand how much tech will (or won’t) be a part of the new management position.


How “Confident” Are You?

This question is something of a misnomer. What I really mean is how are your soft skills and do you enjoy leveraging them? Once you drop into management, you’ll still need to be able to talk the talk and periodically walk the walk of a technologist, but your ability to communicate will become much more important. The same can be said of writing, presentation, and financial skills. Taking some time to evaluate what you truly enjoy doing and where your skills lie will make for a much more informed decision on whether you should take the management plunge.


Do Your Thoughts/Ideals/Morals Align With Your Organization?

While it’s possible to make changes affecting your workplace once you reach management, I’d caution you against thinking you can solve everything wrong with the world and steer your company in a completely different direction. Take the time to look at how your organization treats people. Are other managers satisfied in their roles? Do you generally agree with the direction and leadership of your organization? Depending on the type of person you are, this isn’t a make-or-break question. Regardless of your role, it’s often easier and more effective to row in the same direction as your organization.


How Thick Is Your Skin?

This one is simple: when you lead a team, you’re ultimately responsible for the team’s actions and productivity within the organization. On the surface, this seems like a noble charge, and it is. However, there will be days where people just want to yell at whomever is in charge—you. I work within a fantastic, compassionate organization, and there are still days where “The Buck Stops Here” means I’m going to get earfuls from various corners of the org. It goes with the territory. Are you OK finding yourself in this position? It’s a question you’ll want to consider.


From What Do You Derive Value?

I’m not talking about financial reward of moving into management, although hopefully there’s something of a reward. What I’m talking about is how you motivate yourself, where the sense of accomplishment comes from. Basically, what makes you tick?

For me, this was the biggest adjustment in going from an individual contributor role to management. The reward systems can be very different. As an individual contributor, the feedback is often immediate in that, when you achieve or accomplish a task, there’s immediate satisfaction from completion. Many technologists love this aspect of IT—you’re constantly getting things done, you’re a doer. Depending on the size of your org, when you move into management, you’ll be removed from at least some of the day-to-day “doing.” So, where do the rewards come from?


I’ve had this discussion with leaders of varying sizes of organizations and a couple of themes come to the fore across these chats. Successful and satisfied managers talk about the joy from helping develop people. Being an empathetic person and seeing others succeed can become a powerful replacement for that sense of accomplishment. Having a strategic mindset and seeing your place in a larger puzzle can help ease the loss of the day-to-day tactical achievements.


Is Management Right for Me?

Well, nobody can decide if it’s right for you or not. I’ll share one final thought with you, though. At the start of this piece, I said I was wrong about going into management. That was true. In that role, I wasn’t a great manager, and I voluntarily went back to an engineering role. I’ve since made the jump again, in a situation better suited to me and for an organization who helps to develop my skills and wants to see me succeed as a manager. I have a better idea of what to expect and I’m much happier this time around. Whether you decide to make the leap into management or not, know it’s not a one-way street. Whichever way you decide to go, I hope the thoughts here help you on your journey.


I like sharing my thoughts on the softer side of IT. If you’ve found this article interesting, please check out my other articles The Most Important Skill You're Not Hiring For and Know Your IT Department Frenemies (aka Why Can’t We Be Friends?).

Everyone take a deep breath and calm down. The likeliness of a robot taking over your job any time soon is very low. Yes, artificial intelligence (AI) is a growing trend, and machine learning has improved by leaps and bounds. However, the information technology career field is fairly safe, and if anything, AI/machine learning will only make things better for us in the future. However, a few IT jobs already have experienced the impact of AI, and I want to cover those here. Now, take this with a grain of salt, since AI/machine learning technology is fairly young and a lot of the news out there is simply conjecture.


Help Desk/IT Support

Think about the last time you called a support desk. Did your call get answered by a human or a robot? OK, maybe not an actual robot (that would be awesome), but an interactive voice response (IVR) system. How annoying is that? How often do we just start yelling, “Representative... representative... REPRESENTATIVE!” It can take several routes from an IVR system before we get a human who can help us out. This is all too often the situation when we call support or the help desk. Unfortunately (for help desk specialists), AI is only making IVR more efficient. AI enhances the capability of the IVR system to better understand and process human interaction over the phone. With IVR systems configured for automatic speech recognition (ASR), AI essentially eliminates the need for input via the keypad as it can more intelligently process the human voice response.


Data Center Admins

This one hurts because I’ve done a lot of data center admin work and still do some today. The idea of machine learning or AI replacing this job hits close to home. The truth is automation tools are already replacing common tasks data center admins used to carry out daily. Monitoring tools have used AI to improve data analytics pulled from system scans. Back when I started in IT, the general ratio was around one hundred systems to one administrator. With advances in monitoring, virtualization and AI, it’s now closer to one thousand systems to every administrator. While this is great for organizations looking to cut down on OPEX, it’s not great news for administrators.


Adapt or Die

Yeah, maybe that’s a little exaggerated, but it’s not a bad way to think. If you don’t see the technological advances as a hint to adapt, your career likely will die. AI and machine learning are hot topics right now, and there’s no better time to start learning the ins and outs of it and how you can adapt to work with AI instead of becoming obsolete. Understanding how to bridge the gap between humanity and technology can serve you well in the future. One way you can adapt is by learning programming, thereby gaining a better understanding of automation, AI, and machine learning. Maintain your creativity by implementing new ideas and using AI and machine learning to your advantage.


In the end, I don’t believe AI or machine learning will eliminate the need for a human workforce in IT. The human brain is far more adept at storing, processing, and analyzing data than any robot or machine will ever be. The human brain can adapt, learn, and connect with other humans in ways machines can’t. There might be an influx of jobs being taken over by AI, but we’ll always need humans to program the software and design the machines.

In the IT industry, you’ll hear “I’ll sell you a DevOps; how much is it worth?” But the joke’s on you because you can’t sell (or buy) DevOps, as it is, in fact, an intangible entity. It’s a business process combining software development (Dev) and IT management processes (Ops) with the aim of helping teams understand what goes into making and maintaining applications and business processes. All this happens while working as a team to improve the overall performance and stability of said apps and processes rather than “chucking it over the fence” once your department’s piece of the puzzle is finished.


DevOps is often referred to as a journey, and you probably need to pass several milestones before you could consider your company a DevOps house. Several of the major milestones stem from the idea of adopting a blue/green method of deployment, in which you deploy a new version of your code (blue) running alongside the current version (green) and slowly move production traffic over to the new blue deployment while monitoring the application to see if improvements have been made. Once all the traffic is running on the blue version, you can stage the next change on the green environment. If the blue deployment is a detriment to the application, it’s backed out and all traffic reverts to the current green version.


A key part of the above blue/green deployment is a methodology of continuous integration and continuous deployment (CI/CD), whereby minor improvements are always being undertaken with the goal of optimizing the software and the hardware it runs on. To get to this point you need to make sure you have a system in place to continuously deploy to production, as well as a platform for continual testing. Your QA processes need to tackle everything from user integration to vulnerability testing and change management, and since you don’t want to have to be hunting around finding IP addresses or resource pools to run it on, automation is going to be key.


As you move towards CI/CD adoption rather than separate coding and testing phases, you begin to test as the code is being written. In turn, you’ll start to automate this testing and eventual movement into production, which is referred to as a deployment pipeline. Finally, you’ll also need a more detailed way of performance monitoring, hardware monitoring, software monitoring, and logging. With performance monitoring, it’s no longer good enough to look at network latency—you need to have a way to understand the performance process, including the IO to an application stack, the amount of code commits and bugs identified, the vulnerabilities being handled, and the environment’s health status. With so many moving parts, you’ll also need something to ingest the logs and give you greater insights and analysis to your environment.


But for all this to be undertaken, the first and possibly most major hurdle you’ll have to clear is the cultural shift within the organization. Willingness to cooperate truthfully and honestly as well as making failure less expensive is at the core of this shift. This cultural move must be led from the top down within the company. Making IT ops, software development, and security stop pointing the finger at each other and understand they all have a shared responsibility in the other departments’ undertaking can be a challenge, but if they’re properly incentivized and understand the overall goal, this shift can be a smoother process for an organization.


This building of the correct foundation as per the above milestones allows you thus to move from getting started into the five stages of DevOps evolution: Normalization, Standardization, Expansion, Automated Infrastructure Delivery, and Self-Service. Companies moving into the Normalization stage adhere to true agile methods, and the speed at which they invoke changes begins to increase, so with time they’re no longer hanging around like a loris, taking days or weeks to patch critical vulnerabilities, but move and adapt with the speed of a peregrine falcon.


In the recent Puppet 2019 State of DevOps report, they try to raise the idea of improving your security stance by moving through the five stages of evolution so you can adapt quickly to vulnerabilities. For instance, about 7% of those surveyed can respond within an hour. Those organizations with fully integrated security practices have the highest levels of DevOps evolution. This evolution, in turn, will let you soar through the clouds.


Meh, CapEx

Posted by ian0x0r Nov 7, 2019

Do you remember the good old days when you saved up your hard-earned cash to buy the shiny thing you’d been longing for? You know, like that Betamax player, the TV with the lovely wood veneer surround, or a Chevrolet Corvair? What happened to those days?


I remember as a young lad growing up in the U.K., there were shops like Radio Rentals or Rumbelows where you could get something “On Tick.” ¹ It didn’t seem like the done thing at the time; it almost seemed like a dirty word. “Hey, have you heard, Sheila got her fridge-freezer on tick.” The look of shock on people’s faces!


Now fast forward 25 years and here we are—almost everything is available on a buy now, pay later or rental model. Want a way to access films quickly? Here’s a Netflix subscription. Need something to watch them on? Hey, here’s a 100” Ultra HD 8K, curved, HDR, smart TV with built-in Freeview and Freesat, yours for 60 low monthly payments of £200. Need a new set of wheels? No problem. Something like 82% of all new cars in the U.K. are purchased with a PCP payment plan. This is people reaching for the shiny thing and being able to get it when previously they couldn’t or shouldn’t.


The OpEx Model


So, what’s my point and how is this relevant to the IT landscape?


First and foremost, rental models can work out more profitably for the vendor selling the goods. Slap a little interest premium on the payments as the cost is spread over a longer term. A rental income is more predictable as well, the annuity is the name of the game.


Incentivizing the rental model makes it look more attractive. Look at the cost of an iPhone, for example. Each new release gets more and more expensive. My first four cars combined cost less than a new iPhone 11 Pro. But Apple (and others) are smart. Pay your monthly sum of cash and every 12 months, they’ll give you a new phone. And oh heck, if you drop it and smash the screen, no problem—it’s covered in the monthly cost. You don’t get that service if you buy an iPhone upfront with your hard-earned cash. It makes customers sticky, as in they’re less likely to move elsewhere.


Now let’s look at IT vendors. Microsoft, Amazon, Dell, HPE, you name it, their rental model fits nicely into an OpEx purchasing plan.


There’s no denying Microsoft has killed it with Office 365. This is something I’ll dive deeper into in an upcoming blog post. AWS and all public clouds allow you to pay for what you consume, on a monthly basis, although this cost isn’t always predictable. 


Even hardware vendors are at it. Dell can wrap up your purchase in a nice finance package spread over 3 – 5 years. HPE has their Green Lake program, which is, and I quote, “Delivering the As A Service experience. We live in a consumption-based world—music, TV shows, groceries, air travel and much more. Why? Because consuming services delivers a better outcome faster, in our personal lives and in business. It’s true for music, and it’s also true for IT.” ²




So why Meh, CapEx? In an increasingly diverse IT landscape, with more and more solutions delivered As-A-Service, coupled with the increased pace of innovation, it can be difficult to predict costs and get it right with a CapEx investment lasting 3 – 5 years. I mean, who wants to still be rocking an iPhone 11 in 2024?


¹ On Tick – To pay for something later, via Urban Dictionary https://www.urbandictionary.com/define.php?term=get%20it%20on%20tick

² HPE Green Lake Quote https://www.hpe.com/uk/en/services/it-consumption.html

It’s hot in Dubai. I mean for-real hot, and I’m a native Texan. It’s blistering, cloudless, equatorial sun Northern Hemispher-ians can’t imagine, plus the four-nines humidity of the Mississippi Gulf Coast. Mid-day it’s as shocking as a windy February night in Chicago. It hits you in the face and chest when you walk outside. The unintended effect of which isunlike at US and EMEA conferences where attendees have multiple distractions to considerat GITEX, 100,000 humans were happy to be together for a week. The A/C was glorious. Perhaps the communal relief of the indoors was a foundation for great conversationsand several surprisesat my first conference in the Middle East.


Friends who’ve spent time in the Gulf essentially said, “PFFT. Dubai is the Las Vegas of the region.” And in some ways, that’s true. It’s a convention city, a showplace, and has similar attractions, just turned up to 11. But, while it’s also westerner-friendly, it doesn’t dump everyone into the same cacophonous strip of mostly Nerfed vice and titillation as a gimmick. Instead, it finds a way to accommodate multiple cultures in close proximity. Although, you might wave off the pulled pork tacos at the Tex-Mex place at Le Meridian, especially if you’re from Austin.


It’s likely Dubai is requirements-based rather than organically developed, i.e., it’s intentional and planned like (good) IT. That may be what lead to the second of my surprisesIT professionals attending GITEX were more like US IT pros than some other regions I’ve visited. Typically, the dress and conversation at non-US events is more formal and led by managers instead of geeks. But at GITEX, attendees are the actual engineers who manage data centers, deploy and troubleshoot apps, and who are trying to get a handle on evolving cloud monitoring. But they’re also warm and gracious, patiently restating complex questions to this American, and taking the time to ensure English comms didn’t obscure nuance and technical detail. That’s not easy with acronym- and buzzword-overloaded technology, and it reiterated how eager they were to make real changes to operations, not just play with gear.


It was also different in that while most US SolarWinds customers work with us directly, international customers often work through our partners. GITEX was my first show with two of our regional distributors in the booth with us, Clever Distribution and Spire Solutions. I’ve been fortunate to speak at partner events, but this was my first chance to watch lines of attendees visit and say hi to the people they trust to help select tools for their businesses. Apparently smiles, handshakes, and selfies aren’t only for the THWACK® community at Cisco or Microsoft conferencesour partners have their own active social cosmos. It also let me sip more lattes and answer more “impossible” questions, the answers to which are usually, “oh, that’s been in there since version 12.x, can I show you how to get started?” and “when’s the last time you updated? Did you see the new upgrade tool? (Hint-hint, What We’re Working On, Upgrade Planning, ahem).


But I think what’s stayed with me most, were the number of displays featuring not the just new tech, but new tech in production improving human experience. In one case, tech was even touching hearts. There was med-IoT on ambulances in the Dubai EMS fleet, recent investments in citizen data security and privacy, lots of Smart City tech, and a big focus on useful 5G without solely focusing on vapor deployment. In fact, the largest and most tradeshow-beautiful hall was dedicated to private-public partnership projects, with row after row of not only cool toys, but well-designed projects. The ratio of useful technology to shiny and new was refreshing.


But one stand packed in among all the IT tech, stood out in an unexpected waya Hajj virtual reality simulator. Saudi Arabia’s Ministry of Hajj and Umrah sponsored a stand for Karachi-based Labbaik VR, who’ve spent years developing a PC-based training system for pilgrims planning to perform their first Hajj or Umrah. But recently, they combined their cloud-scale 360 world data with VR headsets. The result is an immersive 3D experience.


It was the sort of technology a bunch of engineers like us would build for fun, but because of the domain, it’s transcending entertainment to truly touch people who use it. Hajj is the fifth pillar of Islam, and Muslims who are physically and financially able have a duty make the pilgrimage to Mecca at least once in their lifetime. But in reality, not all can make it for many reasons, including cost, health, and political boundaries. Speaking with developers from the Labbaik VR team, I heard what it was like as users transitioned from the PC to the 3D headset version. What had been a helpful and engaging simulation became an emotional experience, especially for those who’d likely never get a chance to make the pilgrimage and fulfill their duty. Joy, wonder, and even tears weren’t uncommon.


Perhaps it’s being in technology a long time, or maybe getting my start in customer service and transportation systems, but as technology and human activity become more and more closely intertwined, what we do as IT pros increasingly matters. Sometimes it seems like we’re consumed by never-ending troubleshooting to trim a few milliseconds off a web transition, stuff yet another 100 VMs into our existing clusters, figure out how we’re going to get a handle on containers with no orchestration budget, and on and on. But apps delivering social services, data analysis identifying and correcting long-held biases, and distributed tech touching or even saving lives is becoming more and more common.


Yes, we still must do All the Enterprise Things, but it’s never been a more fulfilling time to pay your mortgage with what we do at our keyboards. For those in Dubai, thanks for coming by the stand. It was, as always, a pleasure to speak in person and not just on THWACK. But more than that, thanks for the stories you shared about your IT adventures throughout the Middle East and Africa. The more I travel around the world for SolarWinds the more I’m in awe of the breadth of work you all do. If tech is the universal language, THWACK members are particularly fluent.

Omar Rafik, SolarWinds Senior Manager, Federal Sales Engineering


Here’s an interesting article by my colleague Brandon Shopp about improving systems management for Microsoft environments. He explores a range of considerations.


Microsoft offers some of its own monitoring options, such as System Center and Windows Admin Center. Federal IT pros can optimize performance by including additional monitoring strategies such as monitoring Windows servers, Microsoft applications, databases, Hyper-V, Azure, and Office 365.


Monitoring Strategies


Windows Servers

Identifying a performance issue involves understanding what’s not operating efficiently. In a Microsoft environment, this means knowing the operating system isn’t part of the problem.


To gain this knowledge, consider tools capable of focusing on the Windows servers to provide highly-specific information and help pinpoint—or rule out—a server-based issue.


Microsoft Applications

It can be impossible to truly understand application health—and, in turn, performance—without understanding how well Microsoft application services, processes, and components are operating.


To get this critical information, consider a tool that gives the federal IT team the ability to:


•Isolate page-load speeds based on location, application components, or underlying server infrastructure

•Monitor requests per second, throughput, and request wait time

•Identify the root cause of problems by monitoring key performance metrics, including request wait time and SQL query executing time

•Identify which webpage elements are slow and affect overall webpage application performance


A greater understanding of the performance levels of the processes feeding in to and out of applications can prove invaluable when trying to identify higher-level application performance issues.



Every federal IT pro knows monitoring database performance is a must.


Specifically, be sure to invest in a tool with the ability to troubleshoot performance problems in real-time and historically. The historical perspective will allow the team to identify a baseline, so they can better understand the severity of a slowdown. This perspective will then allow the ability to analyze the database workload to identify inefficiencies. Ideally, the tool of choice will also provide SQL Server index recommendations as well as alerting and reporting capabilities.



For optimized virtual infrastructure performance, be sure to optimize Microsoft Hyper-V—the company’s virtualization platform.


One of the best ways to do this is by understanding and optimizing the size of virtual machines through capacity planning. It’s also possible to take this even further by predicting the behavior of the virtual environment and solving potential issues before they escalate.


Not all tools will provide these capabilities, so choose wisely.



Many federal IT pros believe cloud monitoring is in the hands of the cloud provider. Not so. It’s possible—and highly recommended—to monitor the cloud infrastructure and transit to help ensure optimized system and application performance.


For example, a good tool will provide the ability to monitor Azure-based applications with as much visibility as on-premises applications. A better tool will go even further and allow the federal IT pro to measure the performance of each network node inside the cloud and to analyze historical performance data to pinpoint a timeframe if performance has degraded.


Microsoft offers a tool called Azure Monitor, which allows the federal IT pro to collect performance and utilization data, activity and diagnostics logs, and notifications from various Azure resources. Azure Monitor integrates with other analytics and monitoring tools, which is a plus for larger environments supporting a range of different types of products and services from a range of vendors.


For further peace of mind—and to help protect against data loss—look for the ability to back up emails to a secondary location.




Operating in a Microsoft-centric world doesn’t mean the federal IT pro must rely only on Microsoft products and services to help optimize performance. Yes, Microsoft has excellent options. But more out there can go a long way toward ensuring a top-performance environment on site or in the Azure cloud.


Find the full article on our partner DLT’s blog Technically Speaking.


The SolarWinds trademarks, service marks, and logos are the exclusive property of SolarWinds Worldwide, LLC or its affiliates. All other trademarks are the property of their respective owners.

We come to you today with a late-breaking breaking postscript! When we last left our IT hero, they were asking themselves deep and probing questions like Will I Save Money by Moving to the Cloud and I’m Still On-Premises, Is That OK? As they go about modernizing their IT organization, they cry out, “What if IaaS isn’t not for me?” Never fear—SaaS is here!


As we laid out in our previous reports from the data center, a SaaS solution is broadly defined as:

  • SaaS (Software-as-a-Service) You subscribe to, access, and consume software that lives and is managed elsewhere.


What Does It Really Mean?

Do you like running an email server? A sales platform? Chat? Antivirus? Security software? File sharing, help desk? The list goes on and on. Let’s cut to the chase. With a SaaS platform, you get to offload many of the traditional management aspects of IT management (is the solution up? Does it need more space? How do I upgrade it?) to someone else and take advantage of all the benefits.


Seems simple enough, right? And maybe it is, but as with any other solution or architecture, you need to ask yourself a few questions. This analysis is especially important if you work in the SMB space as your resources are almost certainly constrained in some way, whether that be people, time, money, leadership, etc. If an enterprise wants to test the water on something like a SaaS solution, where they spend several million dollars and it just doesn’t work out, the decision may end up being immaterial. If you spend weeks trying out a solution (any solution!) while spending tens of thousands, or even hundreds of dollars in some cases, and the solution tanks, the vibrancy of your organization can be negatively impacted.


OK, so you want to give this SaaS thing a chance. That SaaS platform sure looks appealing at first glance, but what questions do you need to ask within your org to help ensure success when you’re looking to adopt a new cloud offering?


  • Efficiencies. As always, what’s driving you to look at a cloud-based offering? Are your people driving sales or A/P off a spreadsheet or perhaps spending hours swivel-chairing data from one place into another? In short, can you derive value from simplifying your activities and increasing efficiency? Yes? Congrats—you’re ready to continue exploring SaaS solutions for your business.


  • Costs. People are expensive. Software is expensive. People managing software is EXPENSIVE! With Software as a Service, you can kick the non-value add parts of the equation out the door and just consume the solution.


  • What do you want to work on? At first glance, this seems very similar to the above, but with a slightly different spin. This speaks to the heart of your organization and your mission. If you run a custom cabinetry business, do you really care where your CRM lives? Of course not. You just want to make some fine woodwork, and you could care less how the nuance of your application is architected as long as you can keep track of your pipeline, materials, billing, and receivables. In circumstances like this, as long as you find the right provider, you’re likely to derive the biggest benefit from a SaaS.


So, after asking yourself questions like this and looking at what makes your organization unique, SaaS still sounds appealing to you. But it’s not all lilacs in the springtime. As with anything in this world, there are trade-offs. Let’s quickly tick through a couple of the typical stumbling blocks for IT organizations looking at SaaS solutions.


  • Connectivity. I’ve seen SMBs run their business off consumer-grade cable connections. Not everyone can have carrier redundancy and multiple legs of dark fiber. If you’re a small shop and you’re going to heavily consume a SaaS application (or anything) requiring lots of data, you need to look at your connectivity. And if you realize you don’t have the network capacity, that needs to roll up into your TCO and ROI calculations to determine if the solution still makes fiscal sense.  If you’d like to dive deeper into connectivity questions and how they impact your cloud strategy, Thwack MVP rschroeder has provided some excellent thoughts on the matter on a previous post. Thank you rschroeder!


  • Lock-in. This really isn’t any different from a traditional on-premises solution, but needs consideration nonetheless. Are you OK hitching yourself to a SaaS provider, potentially for the long haul? Where other type of cloud offerings may provide an easier way to transition between providers, moving from one SaaS platform to another is likely to involve considerable people-hours and dollars. When asking yourself if you want to be married to a SaaS provider for the long haul, make sure you evaluate their strategy, motivations, and fiscal health to ensure a long and happy relationship. Again, this is no different from evaluating an on-premises solution, but since you’re going to be spending vital organization dollars, the evaluation needs to be completed regardless of direction.


  • Loss of control. This can take several forms. When you partner up with a SaaS provider, you’ll likely lose ability to control maintenance windows, patching, upgrades, and so on. Are you OK if the API you use suddenly changes parameters? These are solvable problems, but you need to evaluate your organization’s unique tolerance for some loss of control.


In our last cloud exploration, we asked the question “I’m Still On-Premises, Is That OK?” Yes, it is. If you’re feeling a little bit left behind, a SaaS platform can be an easy way to dip your toes into the cloud. By simplifying how you manage and approach commodity services, with a SaaS solution, you may find your IT superhero has transformed your business in ways unimaginable just a few years earlier.

Filter Blog

By date: By tag:

SolarWinds uses cookies on its websites to make your online experience easier and better. By using our website, you consent to our use of cookies. For more information on cookies, see our cookie policy.