Run 1,000 Docker Redis Containers in Less Than 15 Minutes on a Cluster of 5 Cloud Servers with 2GB of Memory Each
// DCHQ.io | Linux Containers | Application Deployment
While application portability (i.e. being able to run the same application on any Linux host) is still the leading driver for the adoption of Linux Containers, another key advantages is being able to optimize server utilization so that you can use every bit of compute. Of course, for upstream environments, like PROD, you may still want to dedicate more than enough CPU & Memory for your workload – but in DEV/TEST environments, which typically represent the majority of compute resource consumption in an organization, optimizing server utilization can lead to significant cost savings.
This all sounds good on paper — but DevOps engineers and infrastructure operators still struggle with the following questions:
- How can I group servers across different clouds into clusters that map to business groups, development teams, or application projects?
- How do I monitor these clusters and get insight into the resource consumption by different groups or users?
- How do I set up networking across servers in a cluster so that containers across multiple hosts can communicate with each other?
- How do I define my own capacity-based placement policy so that I can use every bit of compute in a cluster?
- How can I automatically scale out the cluster to meet the demands of the developers for new container-based application deployments?
DCHQ, available in hosted and on-premise versions, addresses all of these challenges and provides the most advanced infrastructure provisioning, auto-scaling, clustering and placement policies for infrastructure operators or DevOps engineers.
- A user can register any Linux host running anywhere by running an auto-generated script to install the DCHQ agent, along with Docker and the software-defined networking layer (optional). This task can be automated programmatically using our REST API’s for creating “Docker Servers” (https://dchq.readme.io/docs/dockerservers)
- Alternatively, DCHQ integrates with 13 cloud providers, allowing users to automatically spin up virtual infrastructure on vSphere, OpenStack, CloudStack, Amazon Elastic Cloud Computing, Google Compute Engine, Rackspace, DigitalOcean, SoftLayer, Microsoft Azure, and many others.
This talk focuses on DevOps strategies to assist development of Docker applications directly on CoreOS hosts. James Russell, a DevOps engineer at PlayStation, covers many areas including the installation of customized toolboxes, integration with IDEs, and enabling collaboration with colleagues.
via Tutorial: Create Server with CenturyLink Cloud Java SDK – CenturyLink Cloud Developer Center.
At CenturyLink Cloud, our culture incorporates a DevOps mindset where each team owns the services they develop from implementation to production. In order to apply the DevOps mindset we adhere to these enabling principles: Automate, Self-Service, and Programmable. As an extension of these principles, we also provide public APIs and SDKs for multiple programming languages to enable our customers to also automate, self-serve and program infrastructure and service interactions.
In this tutorial we’ll take a look at the CenturyLink Cloud Java SDK and how we can use it to create a server using a CentOS 6 operating system template in a specific server group in a US East data center. For scripting in the Java Runtime Environment I tend towards using the Groovy programming language and Gradle for build automation. In order to execute the build commands later in this tutorial you must install Gradle so it can be used from your terminal.
Memory leak detection
CPU, HEAP and memory profiling
Root cause analysis
Compiled from source: Strong Loop
via Migrate a VMware Virtual Machine Running Linux to ProfitBricks | ProfitBricks DevOps Central.
This tutorial is designed to assist in moving a virtual machine (VM) running Linux under the VMware hypervisor into the ProfitBricks cloud.
Table of Contents
via How to set up a system status page of your infrastructure – Xmodulo.
If you are a system administrator who is responsible for critical IT infrastructure or services of your organization, you will understand the importance of effective communication in your day-to-day tasks. Suppose your production storage server is on fire. You want your entire team on the same page in order to resolve the issue as fast as you can. While you are at it, you don’t want half of all users contacting you asking why they cannot access their documents. When a scheduled maintenance is coming up, you want to notify interested parties of the event ahead of the schedule, so that unnecessary support tickets can be avoided.
All these require some sort of streamlined communication channel between you, your team and people you serve. One way to achieve that is to maintain a centralized system status page, where the detail of downtime incidents, progress updates and maintenance schedules are reported and chronicled. That way, you can minimize unnecessary distractions during downtime, and also have any interested party informed and opt-in for any status update.
One good open-source, self-hosted system status page solution is Cachet. In this tutorial, I am going to describe how to set up a self-hosted system status page using Cachet.