#1 History of Data Centric Architecture

Subscribe to get the latest

on Thu May 07 2020 17:00:00 GMT-0700 (Pacific Daylight Time)

with Darren W Pulsipher,

In this episode, Darren talks about the history of applications and how recent changes, primarily due to the onslaught of data from the Internet of Things, is affecting data-centric architectures. The infrastructure is ready, but we don%92t yet have a suitable way to manage all our data. There are three elements that need to change to facilitate this process: people (organization), process (operation), and architecture (technology). Darren focuses on the architecture where data and compute are spread over thousands of edge devices and across public and private clouds.


Keywords

#dataarchitecture #softwaredeveloper #microservice #container #virtualization #technology #compute #data

Listen Here


In this episode, Darren talks about the history of applications and how recent changes, primarily due to the onslaught of data from the Internet of Things, is affecting data-centric architectures. The infrastructure is ready, but we don’t yet have a suitable way to manage all our data. There are three elements that need to change to facilitate this process: people (organization), process (operation), and architecture (technology). Darren focuses on the architecture where data and compute are spread over thousands of edge devices and across public and private clouds.

Purpose Built Hardware-software Stacks

How we deploy applications for missions today has not changed significantly in thirty years. A reference architecture that has an application and application stack is built on specific hardware, with compute and storage connected to a network. This model worked well for a long time, and in fact, about a quarter of applications are still being deployed on purpose-built hardware, but it is not optimal today. Technology moves too fast for this model; drifts happen. In addition, there are long development times, high costs, limited re-use of the technology, and lack of integration with other applications.

Virtualization Architectures

About 20 – 25 years ago, hardware virtualization began to resolve some of these issues with the ability to deploy multiple applications on one machine. Applications were no longer tied to specific hardware. Instead of buying five smaller machines, one larger piece of hardware could be used, not just for compute, but for virtual storage and network functions as well, leading to greater cost-effectiveness. As with any development, this raised some new issues: increased security concerns and “noisy neighbors,” meaning one application interfering with the performance of another because of using up IO bandwidth, network, or storage, etc..

Cloud Architectures

In the early 2000s, cloud technology took off. We could now share across multiple organizations. Where virtualization created abstraction of hardware, cloud technology created abstraction of operations, making it easier to manage multiple machines. The cloud architectural idea created “software defined infrastructure,” which makes it easier to spin up and spin down compute, storage and network resources. Other benefits are decreased CapEx and OpEx costs, due to less hardware and manpower. It also gave bursting ability, for example, for retailers during the busy holiday season or the government during the census. With the progression of this technology, the issues of security and noisy neighbors increased because of multiple tenants on the same machine. Another concern is integration costs between public and private clouds. Even with these worries, however, the pros far outweigh the cons in most cases.

Service and Container Architectures

Over the last five to six years, we’ve seen the reinvention of an old technology: containerization. Docker created an easier way to use the previously clunky and difficult-to-use container technology. Application developers, in particular, embraced this technology because it is consistent across multiple environments. The service management layer with containerization of applications and microservices is more application-centric and maps those applications to generic, virtualized hardware that’s been abstracted away. We now have automatic deployment across multiple clouds, we’ve optimized OpEx and also CapEx on the application stack and service stack layer. Fault tolerance is automated, and it’s much easier to integrate with overlay networks, integrate across multiple clouds, create firewalls, do micro-segmentation etc… all via software.

Security, however, is a primary concern. Since containers are easy to deploy across multiple environments, it’s important to focus on security that is “built-in” to the deployment. Also, there is an increase in complexity. Here, we’ve moved away from a three-tier architecture to a multi-tier or even a micro-service architecture with dozens of services working together. Another problem is where and how the data is stored and managed. On the service management layer, storage is a generic container, which doesn’t manage the data itself.

Internet of Things Architectures

Now, when Internet of Things (IoT) is added to this ecosystem, the increased volume of data is spread across hundreds or thousands of smart devices. It also increases visibility, but this creates additional security concerns. Many edge devices are accessible by the public. For example, someone could tamper with a smart city light, a smart traffic signal, a drone, or security camera. The complexity of the different devices, their number and locations, along with the immense amount of data, is enormous.

Data and Information Management Architectures

How do we work through these issues? Organizations are already changing to handle this complexity with new organizations and positions foxing on data management use cases. Previously, there was no place for these use cases to be managed, so we’ve created a new key layer called the distributed information management layer. This layer manages data across all of the IoT, Legacy, and public and private clouds. It matches the data with application stacks and service stacks, so we can dynamically allocate services and applications close to the data, or visa versa. Regulations and sheer size of data can limit the ability to move data to central locations, as we have done traditionally. With this new architecture, several modes of operations can be utilized including disaggregated analytics, data movement and application movement.

Once again, with this expanded architecture, security is of utmost importance. Security needs to run as a common aspect through all the layers. Identity security, meaning access, authorization, and authentication of individuals, IOT devices, applications, services, and even of data is key. Management of identity includes encryption for trusted data and devices.

Conclusion

All of this architecture together is called the Edgemere architecture. Many of the parts already exist; we need to optimize how they work together. The newest area is the distributed information management layer (DIML). Fortunately, we are beginning to see start-ups and more established companies building out the use cases and the architectural elements in this layer.

The Edgemere architecture helps identify the key moving parts of a modern digitally transformed system and how they all fit together.

Intel fits into this ecosystem by providing the key element of a common physical layer to control and manage all of your resources, whether it’s an IOT device, in the data center, or in a remote location. We make it possible for you to efficiently move the data, store it effectively, and process everything. Whether it’s the Xeon processors at the high end, or whether it’s doing inference or AI on the edge at very low power, Intel has a full stack of physical hardware.

Podcast Transcript