The ‘Intergalactic Computer Network’
The underlying concept of cloud computing goes back to the 1960s and the concept of an “Intergalactic Computer Network” by Joseph Carl Robnett Licklider, known as Lick. LIck had an unusual background of psychologist and computer scientist and was referred to as computing’s Johnny Appleseed. His concept of an “Intergalactic Computer Network” at the Advanced Research Projects Agency (ARPA) created by Eisenhower, became the basis for ARPANET, which became the Internet.
It is interesting to note that the frame of reference for this was the mainframe, which was the prevalent form of Enterprise Computing at the time. Ironically, in many ways cloud computing actually evolved from core concepts that are very mainframe centric. By the way, it should come as some surprise that the mainframe, which was called dead back in the 1980s, is growing at 20 percent year over year according to IBM’s latest financials [Disclosure: IBM is a client of the writer].
However Licklider’s vision went well beyond the initial Internet, which was more about communication. This vision was for everyone on the globe to be interconnected and able to access programs and data at any site from anywhere. This is basically the heart of cloud computing, an old visionary concept made real this century. Since those early days the Internet evolved towards this vision with the most recent evolution being Web 2.0.
Birth of the cloud
Likely the biggest step towards the current concept we call cloud computing was the birth of Salesforce.com in 1999. This showcased that the concept of delivering enterprise applications from the Internet then, the cloud now, from a simple website and it was a huge success.
While most of the traditional firms watched Salesforce.com with interest it wasn’t until the launch of Amazon Web Services, AWS, in 2002 that it became clear we were in the midst of major change and things got scary. Up until then a large number of firms had attempted to scale out their IT services ‘as-a-service’ and failed spectacularly, the last and perhaps largest until Amazon was Intel. But Amazon didn’t fail, they were massively successful. Amazon’s services included storage, computation and even human intelligence through and interesting service called Amazon Mechanical Turk. (Basically a tightly targeted temp service for remote workers on the Web).
The cloud gets teeth
AWS continued to evolve until 2006 when it launched its Elastic Compute cloud, which allowed small companies and individuals to rent computer space where they could run their applications. Amazon set the bar as the first widely accessible cloud computing service.
But they weren’t alone under the Web 2.0 concept. In 2009, Google along with others launched a variety of browser-based apps. Google Apps, which targeted Microsoft Office, was perhaps the most memorable. Suddenly individuals were doing things “in the cloud” that were only done on local computers before. This step is considered one of the most important and it drove Microsoft to respond with services like Office 365 and One Drive, massively spreading this change.
Assisting behind the scenes was the concept of virtualization provided by firms like VMware and Microsoft [Disclosure: Microsoft is a client of the writer], which allowed single servers to not only host multiple application loads but to be able to shift those loads dynamically and in real-time allowing these cloud implementations to both scale and to become increasingly less expensive. This rapid rise of capability and reduction in cost is largely what continues to drive cloud computing today.
The dark underbelly of the cloud emerges
However, the cloud isn’t perfect, a service driven largely on value and rapid scale-out has to give something up and that something is often security and reliability, both of which generally remain stronger when hardware is under IT’s direct control. Outages and breaches in a variety of cloud service providers coupled with laws and compliance rules that limit where data can flow (certain types of data can’t flow across boarders or needs extra protection as a result of regulatory requirements).