Most non-IT people will have heard about the word Terraform, will automatically think of this, changing dead planets into Earth like paradises.
But for those of us that work in Infrastructure and Cloud; Terraform is a language definition that allows the deployment of infrastructure as code. There is a loose analogy there, with Terraform you are building your environment exactly as you wish with a pre-configured script to create a predefined end point, just like the science fiction future environment building a new earth from barren rock.
Anybody who has utilized AWS will be aware of Terraform. It was written by Hashicorp and it is one of the primary methods used to automatically build AWS environments in this bold and new DevOps world.
What not a lot of people aware of is that Hashicorp’s Terraform can be utilized to build any infrastructure; all it requires is a provider. Currently Terraform is available for integration into all the major public cloud providers (AWS, Azure, GCP, Oracle Cloud and Alibaba Cloud). It is also available for on-prem environments with VMware and Microsoft Hyper-V.
Docker, the leading vendor in the container industry, has been dipping into its pockets again, this time to acquire French file sync company Infinit. Yep, Docker, the company whose core product is a container management and deployment technology, is acquiring a consumer file sync company.
At first glance, this is an odd acquisition. Infinit is a company with a product that appears similar to Aspera, Accellion, or even good old-fashioned FTP. If you need to share your files, but email send or receive limits are getting in the way, Infinit allows you to create a Share directly to other Infinit users or a link that can be emailed to non–Infinit users to access your files directly, thereby allowing them to download them for review or further work.
t’s the end of the year, and a good time for thinking back. I’m thinking back to a dark past long ago, when physical servers ran server operating systems, and ran applications—when those servers plugged into a switch, and each endpoint was a single server. The network team could see every device, endpoint, or switch, and could trace packets from end to end. Network admins would tell you that those were Golden Days, when troubleshooting was easy and networks were simple. Then, ten or so years ago, along came server virtualization. All of a sudden there were multiple servers on any given endpoint, and worse, the servers would move between endpoints not only at will, but mid-flow. Troubleshooting became Hard, with a capital H.
Out of this came innovations such as VMware’s dvSwitch and the Cisco 1000V distributed vSwitch. These gave network admins the tools they required to push their traces deeper into the virtualization systems and to regain the end-to-end connectivity they desired. As time progressed, the ability to mirror flows and to extend technologies such as NetFlow into the hypervisor brought the VM world back into network admins’ view. As time advanced further, network functions virtualization (NFV) moved some of the functions of the network into the hypervisor, or into VMs, but the interaction between the flows remained fairly constant. The more recent developments of overlay/underlay networks have again pushed the end-to-end traffic flows into the twilight of tunnels (encrypted or not). The two-tier network model has made troubleshooting harder again, with layer 2 networks tunneled through layer 3 switch interconnects. Now Docker is throwing another spanner in the works.