TIME-BOUND LIVE MIGRATION WITH MINIMAL STOP-AND-COPY

- Oracle

A computing system may access a source device configured to store device traffic in a source device queue. The computing system may launch, by one or more worker nodes of a control plane of the computing device, a destination device. The computing system may transmit a source device state to the destination device. The control plane worker nodes may configure the destination device from the source device state. The computing system may instruct the source device to pause storing the device traffic to the source device guest memory. The computing system may store dirty page information in the computing system memory. The computing system may transmit the dirty page information to the destination guest memory. The computing system may instruct the destination device to resume storing the device traffic to the destination guest memory.

Skip to: Description  ·  Claims  · Patent History  ·  Patent History
Description
BACKGROUND

A virtual machine (VM) may be migrated between host machines, but migrating host machines can cause service interruptions for virtual machine end users. If the virtual machine can be migrated within a sufficiently short timeframe, the user may not experience a noticeable interruption. However, copying device state information from a source host to a destination host device can be time consuming. Accordingly, improvements in virtual machine migration are desirable.

BRIEF SUMMARY

A system of one or more computers can be configured to perform particular operations or actions by virtue of having software, firmware, hardware, or a combination of them installed on the system that in operation causes or cause the system to perform the actions. One or more computer programs can be configured to perform particular operations or actions by virtue of including instructions that, when executed by data processing apparatus, cause the apparatus to perform the actions.

In one general aspect, techniques may include accessing a source device configured to store device traffic in a source device guest memory. The source device may be accessed by a virtual machine manager of a computing device. The techniques may include launching a destination device by one or more worker nodes of a control plane of the computing device. The techniques include storing a source device state in a computing system memory. The source device state can be stored by one or more worker nodes of a control plane of the computing device. The techniques may include transmitting the source device state to the destination device. The techniques may also include instructing the source device to pause storing the device traffic to the source device guest memory. The techniques may include storing dirty page information in the computing system memory. The techniques may include transmitting the dirty page information from the computing system memory to a destination device guest memory. The techniques may include instructing the destination device to resume storing the device traffic to the destination device guest memory. The techniques may include methods and corresponding computer systems, apparatus, and computer programs recorded on one or more computer storage devices, each configured to perform the actions of the techniques.

Implementations may include one or more of the following features. Techniques that include: limiting a bandwidth of device traffic to the source device. Techniques where the bandwidth can be limited based at least in part on a queue depth of a source virtual machine (VM) queue or number of path groups. Techniques where the computing system is virtual function device that represents a nonvolatile memory express (NVMe) device. Techniques where a time period between instructing the source device to pause storing device traffic and instructing the destination device to resume storing device traffic is 500 milliseconds (ms) or less. Techniques where the dirty page information includes device traffic that is sent to the source device guest memory while the device traffic is paused. Techniques where the source device state includes namespace object information. Implementations of the described techniques may include hardware, a method or process, or a computer tangible medium.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 shows a simplified diagram of a virtual machine according to an embodiment.

FIG. 2 shows a process for sending traffic from a virtual machine (VM) to a target on a network via a virtual function (VF) according to an embodiment.

FIG. 3 shows a diagram of a source host and a destination host connected to a control plane according to an embodiment.

FIG. 4 shows a process for migrating a virtual machine (VM) from a source host to a destination host according to an embodiment.

FIG. 5 shows a method for performing a live migration according to an embodiment.

FIG. 6 is a simplified control path diagram showing cloud infrastructure components for attaching persistent storage according to an embodiment.

FIG. 7 is a diagram showing a kernel architecture for implementing Internet Small Computer Systems Interface (i SC SI) and Non-Volatile Memory Express (NVMe) attachments according to an embodiment.

FIG. 8 is a Non-Volatile Memory Express (NVMe) system diagram according to an embodiment.

FIG. 9 is a diagram of a Non-Volatile Memory Express (NVMe)/Transmission control Protocol (TCP) target according to an embodiment.

FIG. 10 is a simplified diagram of a smart network interface card (smartNIC) with Non-Volatile Memory Express (NVMe) according to an embodiment.

FIG. 11 is a diagram showing multipath handling in a smart network interface card (smartNIC) according to an embodiment.

FIG. 12 shows a diagram of an architecture for performing encryption/decryption with a smart network interface card (smartNIC) according to an embodiment.

FIG. 13 is a block diagram illustrating one pattern for implementing a cloud infrastructure as a service system, according to at least one embodiment.

FIG. 14 is a block diagram illustrating another pattern for implementing a cloud infrastructure as a service system, according to at least one embodiment.

FIG. 15 is a block diagram illustrating another pattern for implementing a cloud infrastructure as a service system, according to at least one embodiment.

FIG. 16 is a block diagram illustrating another pattern for implementing a cloud infrastructure as a service system, according to at least one embodiment.

FIG. 17 is a block diagram illustrating an example computer system, according to at least one embodiment.

DETAILED DESCRIPTION

In the following description, various embodiments will be described. For purposes of explanation, specific configurations and details are set forth in order to provide a thorough understanding of the embodiments. However, it will also be apparent to one skilled in the art that the embodiments may be practiced without the specific details. Furthermore, well-known features may be omitted or simplified in order not to obscure the embodiment being described.

Live migration can refer to the transfer of virtual machines (VMs) between host devices within a specified time period. The transfer can occur within a set time period so that virtual machine end users do not notice an interruption in service. Live migration can include a pre-copy phase that can prepare virtual functions or hosts for a stop-and-copy phase when the source host machine is stopped, and its contents are copied to the destination host machine. The pre-copy phase can occur before the stop-and-copy phase, and the pre-copy phase can include provisioning a destination host machine. The destination host machine may be provisioned before the pre-copy phase.

The stop-and-copy phase can begin by stopping traffic to the non-volatile memory express (NVMe) devices associated with the virtual machine at the source host. The stop-and-copy phase can include copying the remaining device state and dirty page(s) from the source host to the destination host and resuming traffic to the virtual machine at the destination host or traffic can be resumed after the stop and copy phase has concluded. The stop-and-copy phase may need to occur within a specified time period (e.g., 500 milliseconds) for the host migration to be a live migration. The stop and copy phase, and resuming traffic, may need to occur within the specified time period. The time period can be short enough that the virtual machine's users do not notice an interruption in service.

Before the pre-copy phase, a destination virtual function can be launched on the destination device by the control plane. The control plane can be part of a computing system that determines how packets are routed. The control plane can launch a destination virtual function (VF) on a destination device using an agent called a control plane worker node. The destination device can be a smart network interface card (smartNIC). Namespaces and path groups in a smartNIC associated with the virtual function or virtual machine can be created on the destination device by the control plane worker node.

There are a number of use cases for live migration, and, for example, live migration can allow for host maintenance with minimal disruption to virtual function end users. A virtual function can be moved from a source host requiring maintenance to an available destination host that does not require maintenance. After migration, the source host can receive maintenance without interrupting the virtual function running on the destination host. The maintenance can include firmware upgrades, kernel upgrades or hardware upgrades.

Live migration can be used to change the host resources available for a virtual machine. For example, a virtual function can be moved from a source host to a destination host with more central processing unit (CPU) threads if a virtual function needs more processing power. Live migration can allow for resource optimization and load balancing. For instance, network latency can be reduced by migrating a virtual function to a destination host that is closer to the virtual function's storage. Additionally, virtual functions can be consolidated onto fewer hosts to conserve power, or virtual functions can be spread onto more hosts to make the network more robust to the failure of a host.

In an illustrative example, a client of a cloud service is using a virtual machine provided by a cloud service provider. In this case, the virtual machine is a system virtual machine and the client is using programs on an operating system that is running on a server. Traffic is routed to and from the virtual machine by a virtual function. A virtual machine manager (VMM) coordinates the server's hardware to support virtual machines including the client's virtual machine. The server is due for scheduled maintenance and the server must be shut down.

The cloud service provider wants to perform the scheduled maintenance without the client knowing the server hosting the client's virtual machine has changed. To transfer the virtual machine from a source server, currently hosting the virtual machine, to a destination server without the client noticing, the cloud service provider wants to migrate the virtual machine between servers within a short time period. In this case, the cloud service provider decides to migrate the virtual machine within 500 milliseconds (ms) time period.

To migrate the virtual machine, the cloud service provider begins by instructing the virtual machine manager to launch a destination virtual function on the destination server. Worker nodes of the cloud service provider's control plane copy the source virtual function's state and transfer the state to the destination virtual function. Once the destination virtual function is configured, traffic through the source virtual function is paused and a the 500 ms time period begins.

During the 500 ms time period, the virtual machine manager can transfer the device state, which can include traffic from the source virtual function queue in the source device's guest memory, to the destination virtual function queue in the destination device's guest memory. The virtual machine manager also transfers dirty page(s), from the source virtual function to the destination virtual function. Dirty page(s) can include modified virtual memory pages that have not been written to storage. Once the queue and dirty page(s) have been transferred, the time period can end, the virtual machine manager can resume traffic for the virtual machine through the destination virtual function, and the virtual machine can begin to run on the destination device. The time period may end after the traffic through the virtual function resumes.

FIG. 1 shows a simplified diagram 100 of a virtual machine according to an embodiment. A virtual machine (VM) 105 can be created and run using a hypervisor 110 (e.g., a virtual machine manager (VMM)) such as quick emulator (QEMU). A hypervisor can be software that can create and run one or more virtual machine instances. An operating system (OS) 115 (e.g., Windows, MacOS, Linux, etc.) can run on virtual machine 105. An OS can be software that manages hardware and software applications. One or more software applications, such as applications 120a-120b, can run on operating system 115 within virtual machine 105.

Virtual function (VF) 125 can be a virtualized instance of a nonvolatile memory express (NVMe) device, and virtual function 125 can support single root input/output virtualization (SR-IOV). A network adapter, such as smart network interface card (smartNIC) 130, can support one or more virtual function instances. Physical resources in smart network interface card (smartNIC) 130 can be shared between one or more virtual function instances (e.g., virtual function 125). A network adapter can be hardware that can connect a computer to a network, and a smartNIC (e.g., smartNIC 130) can be a type of network adapter with enough processing power to perform networking functions, security functions, and/or storage functions.

Traffic can travel between a network and virtual machine 105 via a pathway that extends through the kernel 135 to the path groups 140a-d via virtual function 125. Kernel 135 can contain software that facilitates interactions between software and hardware components. Traffic entering or exiting virtual machine 105 can be queued at virtual function input output queue (VFIO) 145. A virtual function input output queue can be a framework that can expose direct device access to a userspace. The virtual function input output peripheral connect express driver (VFIO PCI) 150 can support the creation of virtual functions, and virtual function input output peripheral connect express driver 150 can forward packets between virtual function input output queue 145 and virtual function 125.

Virtual function 125 can forward traffic between hypervisor 110 and smartNIC 130. Traffic can be sent to virtual function 125, or received from virtual function 125, at the non-volatile memory express/peripheral component interconnect express (NVMe/PCIe) controller 155 in the Programming Protocol-Independent Packet Processors (P4) pipeline 160. A network controller, such as NVMe/PCIe controller 155, can be a hardware component that connects a computer to a network. P4 pipeline 160 can be a packet forwarding pipeline that uses the P4 coding language that is optimized for packet forwarding in network devices.

Traffic can be forwarded to a network from NVMe/PCIe controller 155, or received at NVMe/PCIe controller 155 from a network, via a namespace and path group. A virtual function (e.g., virtual function 125) can be associated with up to 32 namespaces (e.g., NVMe namespace 1 165, NVMe namespace 2 170, etc.), and a namespace can have up to 16 path groups (e.g., path groups 140a-140d). However, other configurations with different numbers of namespaces or path groups are contemplated. In addition, the relative amounts of namespaces and path groups can be different across the configurations. A namespace can be a volume that can be accessed by the host. In some embodiments, a namespace can be a partition of computer resources and a process running in a namespace can see and use the resources partitioned to the namespace. A path in one of the path groups 140a-140d can be a single IO connection to a target. Path groups 140a-140d can contain at least one active path and one or more passive paths to allow active-passive failover. The traffic can be forwarded to a target 175 on a network, and target 175 can be a database storing data from virtual machine 105.

A device state for a virtual function, such as a source device state or a destination device state, can include at least one of the configuration of the NVMe/PCIe controller 155 associated with a virtual function (e.g., virtual function 125), the number and configuration of the namespace objects associated with a virtual function (e.g., NVMe namespace 1 165, NVMe namespace 2 170, etc.), or the number and configuration of path groups (e.g., path groups 140a-140d) associated with a virtual function.

FIG. 2 shows a process 200 for correcting dirty page(s) according to an embodiment. This process, in addition to the process from FIG. 4, is illustrated as a logical flow diagram, each operation of which can be implemented in hardware, computer instructions, or a combination thereof. In the context of computer instructions, the operations may represent computer-executable instructions stored on one or more computer-readable storage media that, when executed by one or more processors, perform the recited operations. Generally, computer-executable instructions include routines, programs, objects, components, data structures and the like that perform particular functions or implement particular data types. The orders in which the operations are described are not intended to be construed as a limitation, and any number of the described operations can be combined in any order and/or in parallel to implement the processes or the method.

Dirty page(s) can include modified virtual memory pages that have not been moved to the destination device guest memory. Guest memory can be the memory that is available to the guest operating system running on the virtual machine (e.g., the memory available to OS 115 running on virtual machine 105). Dirty page(s) can occur when a flow of traffic from a virtual machine (VM) 105 to storage (e.g., target 175) has been interrupted. The flow of traffic can be interrupted when smartNIC 130 sends data to the guest (VM) memory via direct memory access (DMA). Turning to process 200 in greater detail, at block 210, a flow of traffic from a virtual machine to a target can be initiated. The traffic can originate from an application 120a-120b running on an operating system (OS) 115. The traffic can be sent from virtual machine 105 to a target 175 via kernel 135 and smartNIC 130. Target 175 can be non-volatile computer storage such as a solid state drive (SSD), hard disk drive (HDD), etc.

At block 220, the flow of traffic can be interrupted. The flow of traffic can be interrupted to migrate virtual machine 105 from a first virtual function (VF) (e.g., virtual function 125) or a first smartNIC (e.g., smartNIC 130) to a second virtual function or a second smartNIC.

At block 230, the dirty page(s) can be identified. In some circumstances, the dirty page(s) can be identified before the flow of traffic is interrupted. The dirty page(s) can be tracked by the smartNIC 130.

At block 240, the dirty page(s) can be forwarded to guest memory. Hypervisor 110 can retrieve the dirty page(s) information from smartNIC 130 and the hypervisor 110 can use the retrieved information to forward the dirty pages from the guest memory available to virtual machine 105 (e.g., the source device) to the guest memory available to a second virtual machine (e.g., a destination device). Traffic can travel from virtual machine 105 to the second virtual machine via a network link (e.g., network link 355 of FIG. 3, described in further detail below).

At block 250, the flow of traffic can resume. The traffic can resume because virtual machine 105 has been migrated from a first virtual function (VF) (e.g., virtual function 125) or a first smartNIC (e.g., smartNIC 130) to a second virtual function or a second smartNIC.

FIG. 3 shows a diagram 300 of a source host and a destination host connected to a control plane according to an embodiment. The control plane 305 can control packet forwarding, and control plane 305 can execute commands via a control plane worker node 310 (e.g., block smartNIC agent (B S A) 660 (FIG. 6)). Control plane worker node 310 can be an agent that can configure namespaces and paths for a virtualized instance of a NVMe device (e.g., virtual function (VF) 125, virtual function 315a-315b, etc).

Control plane worker node 310 can observe the device state for a source smartNIC 345a in source host 320. The device state can include data about namespaces (e.g., NVMe namespace 1 165, NVMe namespace 2 165, NVMe namespace 1 325a-325b, NVMe namespace 2 330a-330b, etc.), path groups (e.g., path groups 140a-140d, path groups 335a-335h, etc.), or paths (e.g., active paths 1180a-1180d (FIG. 11), passive paths 1185a-1180h (FIG. 11), etc.) for source host.

Control plane worker node 310 can use the observed device state from the source smartNIC 345a in source host 320 to configure a destination smartNIC 345b in destination host 340. Configuring the destination smartNIC can include creating a destination virtual function 315b. In some implementations, the device state from the source smartNIC 345a (e.g., P4 Pipeline 330a and virtual function 315a) can be stored in control plane 305. Control plane worker node 310 can use the stored device state to configure the destination smartNIC 345b (e.g., P4 Pipeline 330b and virtual function 315b). Creating or configuring the destination smartNIC 345b can include replicating the namespaces, path groups, and/or paths from source smartNIC 345a in destination smartNIC 345b. The namespaces, path groups, and/or paths can be part of a P4 pipeline (e.g., P4 pipeline 160, P4 pipeline 330a-b, etc.) within a smart network interface card (e.g., smartNIC 345a-b, etc.). Creating or configuring the destination smartNIC 345b can also include creating a NVMe/PCIe Controller 350a-b (e.g., NVMe/PCIe controller 155, etc.). SmartNIC 345a can contain multiple source NVMe devices with one virtual function for each virtual machine.

Source host 320 and destination host 340 can include a hypervisor 355a-355b (e.g., hypervisor 110). When a virtual machine (VM) is migrated from source host 320 to destination host 340, the hypervisors 355a-355b can transfer data from virtual machine 360a or kernel 365a to virtual machine 360b or kernel 365b. The transferred data can include information about an operating system (OS) (e.g., operating system 115, operating system 370a-370b, etc.), information about one or more applications (e.g., applications 120a-120b, applications 375a-375d, etc.), data from the virtual function input/output (VFIO) queue (VFIO 145, VFIO 380a-380b, etc.), the configuration of VFIO PCI (e.g., VFIO PCI 150, VFIO PCI 385a-385b, etc.), etc. This information can be transferred from the source host 320 to the destination host 340 via a network link 355.

FIG. 4 shows a process 400 for migrating a virtual machine (VM) from a source host to a destination host according to an embodiment. Turning to process 400 in greater detail, at block 410, the destination device can be configured. The destination device can be destination smartNIC 345b and the destination device can be associated with a destination virtual machine 360b. The destination device can include a destination virtual function (VF) 315b and a P4 pipeline 330b.

The destination device can be configured by a control plane worker node (e.g, control plane worker node 310, etc.). Configuring the destination device can include transferring a device state from the source smartNIC 345a to the destination smartNIC 345b. The device state can include the namespaces associated with a virtual function (e.g., virtual function 315a), the path groups associated with the namespaces, and/or controller states (e.g., features, asynchronous (asynch) commands, SQ/CQ configuration, etc.).

The device state can be retrieved by or from the control plane 305 and the device state can be passed to the control plane worker node 310 from the control plane 305. The control plane worker node 305 can set up a NVMe/PCIe controller (e.g., NVMe/PCIe controller 155, NVMe/PCIe controller 350a-350b, etc.) and the device state in the P4 pipeline (e.g., P4 pipeline 160, P4 pipeline 330a-330b, etc.). In some circumstances, control plane worker node 310 (e.g., smartNIC agent) may retrieve a volume decryption-encryption key (DEK), and control plane worker node 310 can use the volume DEK to decrypt data received from control plane 305. In some implementations, control plane worker node 310 can retrieve the DEK and pass the DEK to the data plane. The data plane can then use the DEK to encrypt and decrypt the data. Control plane 305 can use hypertext transfer protocol secure (HTTPS) or transport layer security (TLS) for secure communication with the control plane worker node 310

At block 420, traffic to the source device can be stopped. The source device can be source smartNIC 345a and the traffic can be addressed to targets such as target 175, target server 650, or NVMe/TCP target servers 1130a-1130i. When the traffic is stopped, received traffic may not be enqueued to the virtual function input/output queue (VFIO; VFIO 145, VFIO 380a-380b, etc.). The traffic can be stopped by a hypervisor (e.g., hypervisor 110, hypervisor 355a-355b, etc.) running on the source device (e.g., source host 320, etc.). The migration may not be considered a live migration if the traffic is stopped for longer than a time period. The time period can be chosen with a duration that is short enough that someone using the migrated virtual machine may not notice an interruption in service. In some circumstances, the time period can be 500 milliseconds (ms) or less.

At block 430, outstanding traffic can be processed. The pending traffic can include requests in the VFIO queue 380a. Processing the pending traffic can create dirty page(s) because the pending traffic can change entries in the virtual machine memory and the changed entries cannot be saved to guest memory while the traffic is stopped. As discussed above, dirty page(s) can include modified virtual memory pages.

At block 440, the pending traffic can be processed and the dirty page(s) can be copied to the destination device. The dirty page(s) can be copied to the destination device by a hypervisor (e.g., hypervisor 110, hypervisor 355a-355b, etc.). The hypervisor can be a virtual machine manager (VMM) or a machine emulator such as QEMU. In some circumstances, state information can be copied to destination device 340 by a hypervisor. For example, controller states can be copied to the destination device by hypervisor 355a-355b while traffic to source device (e.g., source smartNIC 345a) is stopped, and the rest of the state information can be copied by control plane worker node 310 before traffic to and from source device is stopped.

At block 450, traffic can be resumed. The traffic can be resumed at the destination device. The traffic can be addressed to targets in a virtual machine (VM) 360b that is now receiving traffic via at least one of a virtual function (VF) 315b or a p4 pipeline 330b running on destination device (e.g., destination smartNIC 345b), and the traffic can be enqueued to VFIO 380b.

FIG. 5 shows a method 500 for performing a live migration according to an embodiment. This method is illustrated as a logical flow diagram, each operation of which can be implemented in hardware, computer instructions, or a combination thereof. In the context of computer instructions, the operations may represent computer-executable instructions stored on one or more computer-readable storage media that, when executed by one or more processors, perform the recited operations. Generally, computer-executable instructions include routines, programs, objects, components, data structures and the like that perform particular functions or implement particular data types. The orders in which the operations are described are not intended to be construed as a limitation, and any number of the described operations can be combined in any order and/or in parallel to implement the processes or the method.

At block 510, a source device can be accessed. The source device can be a computing system like the one disclosed in diagram 100, source smartNIC 345a, destination smartNIC 345b, etc. The source device can be accessed by a virtual machine manager (e.g., hypervisor, etc.) of a computing device. The source device can be accessed by the virtual machine manager via a network (e.g., Peripheral Component Interconnect Express (PCIe), the Internet, network link 355, etc.).

At block 520, a destination device can be configured. Configuring the destination device can comprise launching at least one of a virtual function (VF) 125, 315a-315b, or a P4 pipeline 160, 330a-330b on the destination device. The destination device can be configured by the worker nodes of the control plane of a computing device. The destination device can be a computing system like the one disclosed in diagram 100, source smartNIC 345a, destination smartNIC 345b, etc. The destination device and source device can be NVMe devices. The source device can be accessed by the worker nodes via a network (e.g., PCIe, the Internet, network link 355, etc.).

At block 530, a source device state can be stored. The device state can be stored by the worker nodes of the control plane of a computing device. A device state for a virtual function can include at least one of the configuration of the NVMe/PCIe controller 155 associated with a virtual function (e.g., virtual function 125, virtual function 315a-315b, etc.), the number and configuration of the namespace objects associated with a virtual function (e.g., NVMe namespace 1 165, NVMe namespace 2 170, NVMe namespace 1 325a-325b, NVMe namespace 2 330a-b, etc.), or the number and configuration of path groups (e.g., path groups 140a-140d, path groups 335a-355h, etc.) associated with a virtual function. The virtual function can be a virtualized instance of a network adapter.

At block 540, the source device state can be transmitted. The source device state can be transmitted from the source device to the destination device. The source device state can be the memory, path groups and NVMe namespaces for source smartNIC 345a. A worker node of the computing system control plane can transmit the device state. The device state can be transmitted via a network (e.g., the Internet or network link 355).

At block 550, storing traffic can be paused. The virtual machine manager of the computing system can instruct the source device to pause storing traffic. The paused traffic can include traffic stored to a source virtual machine in the source device (e.g., VFIO 145, VFIO 380a-380b, etc.). In some circumstances, the traffic can be paused for 500 milliseconds (ms) or less. In some circumstances, the bandwidth of outstanding traffic source virtual machine can be limited before the traffic is paused. The bandwidth (e.g., outstanding requests/traffic) can be limited so that the amount of source virtual machine dirty data can be transmitted from the source virtual machine to a destination virtual machine (e.g., virtual machine 360b, etc.) within a time limit. The time limit can be 500 ms or less. The outstanding requests/traffic can be limited by the NVMe device based at least in part on the queue depth in the NVMe/TCP queue source virtual machine and number of NVMe/TCP connections.

At block 560, the dirty page(s) can be stored. The dirty page(s) can be stored by the virtual machine manager of the computing device. The dirty page(s) (e.g., dirty pages etc.) can be data source virtual machine (e.g., virtual machine 360a) that is written from the source virtual machine guest memory to the to the destination device guest memory using direct memory access (DMA). The dirty page(s) can be stored to computing system memory (e.g., hard disk drive (HDD), solid state drive (SSD), random access memory (RAM), etc.). The dirty page(s) can be transmitted while storing traffic is paused.

At block 570, the dirty page(s) can be transmitted. The dirty page(s) can be transmitted by the virtual machine manager of the computing system (e.g., source host 320). The dirty page(s) can be transmitted from source guest memory (e.g., the memory available to virtual machine 360a) to destination guest memory (e.g., the memory available to virtual machine 360b). The virtual machine manager of the computing system can transmit the device state from the source NVMe device (e.g., source smartNIC 345a, etc.) to the destination NVMe device (e.g., destination smartNIC 345b, etc.). The device state can be transmitted after traffic is paused.

At block 580, the storing the device traffic can be initiated. The device traffic can be routed by at least one of a virtual function (VF) or a p4 pipeline, and the device traffic can be stored to a destination virtual machine of the destination device (e.g., VFIO 380b, etc.). The device can be resumed by the virtual machine manager of the computing system and the traffic can resume at the destination virtual machine.

As noted above, infrastructure as a service (IaaS) is one particular type of cloud computing. IaaS can be configured to provide virtualized computing resources over a public network (e.g., the Internet). In an IaaS model, a cloud computing provider can host the infrastructure components (e.g., servers, storage devices, network nodes (e.g., hardware), deployment software, platform virtualization (e.g., a hypervisor layer), or the like). In some cases, an IaaS provider may also supply a variety of services to accompany those infrastructure components (e.g., billing, monitoring, logging, load balancing and clustering, etc.). Thus, as these services may be policy-driven, IaaS users may be able to implement policies to drive load balancing to maintain application availability and performance.

In some instances, IaaS customers may access resources and services through a wide area network (WAN), such as the Internet, and can use the cloud provider's services to install the remaining elements of an application stack. For example, the user can log in to the IaaS platform to create virtual machines (VMs), install operating systems (OSs) on each VM, deploy middleware such as databases, create storage buckets for workloads and backups, and even install enterprise software into that VM. Customers can then use the provider's services to perform various functions, including balancing network traffic, troubleshooting application issues, monitoring performance, managing disaster recovery, etc.

In most cases, a cloud computing model will require the participation of a cloud provider. The cloud provider may, but need not be, a third-party service that specializes in providing (e.g., offering, renting, selling) IaaS. An entity might also opt to deploy a private cloud, becoming its own provider of infrastructure services.

In some examples, IaaS deployment is the process of putting a new application, or a new version of an application, onto a prepared application server or the like. It may also include the process of preparing the server (e.g., installing libraries, daemons, etc.). This is often managed by the cloud provider, below the hypervisor layer (e.g., the servers, storage, network hardware, and virtualization). Thus, the customer may be responsible for handling (OS), middleware, and/or application deployment (e.g., on self-service virtual machines (e.g., that can be spun up on demand) or the like.

In some examples, IaaS provisioning may refer to acquiring computers or virtual hosts for use, and even installing needed libraries or services on them. In most cases, deployment does not include provisioning, and the provisioning may need to be performed first.

In some cases, there are two different challenges for IaaS provisioning. First, there is the initial challenge of provisioning the initial set of infrastructure before anything is running. Second, there is the challenge of evolving the existing infrastructure (e.g., adding new services, changing services, removing services, etc.) once everything has been provisioned. In some cases, these two challenges may be addressed by enabling the configuration of the infrastructure to be defined declaratively. In other words, the infrastructure (e.g., what components are needed and how they interact) can be defined by one or more configuration files. Thus, the overall topology of the infrastructure (e.g., what resources depend on which, and how they each work together) can be described declaratively. In some instances, once the topology is defined, a workflow can be generated that creates and/or manages the different components described in the configuration files.

In some examples, an infrastructure may have many interconnected elements. For example, there may be one or more virtual private clouds (VPCs) (e.g., a potentially on-demand pool of configurable and/or shared computing resources), also known as a core network. In some examples, there may also be one or more inbound/outbound traffic group rules provisioned to define how the inbound and/or outbound traffic of the network will be set up and one or more virtual machines (VMs). Other infrastructure elements may also be provisioned, such as a load balancer, a database, or the like. As more and more infrastructure elements are desired and/or added, the infrastructure may incrementally evolve.

In some instances, continuous deployment techniques may be employed to enable deployment of infrastructure code across various virtual computing environments. Additionally, the described techniques can enable infrastructure management within these environments. In some examples, service teams can write code that is desired to be deployed to one or more, but often many, different production environments (e.g., across various different geographic locations, sometimes spanning the entire world). However, in some examples, the infrastructure on which the code will be deployed must first be set up. In some instances, the provisioning can be done manually, a provisioning tool may be utilized to provision the resources, and/or deployment tools may be utilized to deploy the code once the infrastructure is provisioned.

Creating and running a cloud service can include mounting and connecting persistent storage to cloud instances. The persistent storage can be created, using a console or application programming interface (API), and linked to cloud instances (e.g., a virtual machine (VM) host or a bare metal (BM) host running in the cloud). Linking, or attaching, persistent storage to a cloud instance can be performed using a communication protocol. The attached storage can communicate with the cloud instance's guest operating system (OS) using the protocol.

Connections between a cloud instance and persistent storage are flexible and a number of configurations are possible. For instance, the persistent storage can be attached to one or more cloud instances simultaneously. The data in the persistent storage is durable and the storage can retain data after an attachment to a cloud instance is removed. Data can be migrated between instances by detaching persistent storage from one cloud instance and attaching the storage to a second instance.

Durable persistent storage can allow for instance scaling. A cloud instance can be deleted without destroying or reformatting the instance's persistent storage. After the cloud instance is deleted, the instances' persistent storage can be attached to a new instance. The new instance can be created with a different instance type or shape. For example, the new cloud instance can be a VM or a BM regardless of the deleted instance's type. Additionally, the number of cores in a cloud instance can be changed by deleting an initial instance and creating a new instance with a different number of cores.

A transfer of data through an attachment can be started with an endpoint called an initiator. Data can be sent from the initiator to an endpoint that can receive data called a target. An agent can set up the target to receive data and forward the data to the target. A number of advantages can be provided by locating the initiator in a smart network interface card (smartNIC). A user may need to provide login information or other configuration from the cloud instance if the initiator is located in the instance. Additionally, it can be difficult to keep the initiator functional across different guest OS types and OS versions. Locating the initiator in the smartNIC can also free customer resources that would be used to run the initiator.

Attachments can be provided using storage networking standards including Internet Small Computer Systems Interface (iSCSI), paravirtualized (PV) iSCSI, and Non-Volatile Memory Express (NVMe). iSCSI can provide attachments for bare metal (BM) devices with the initiator running from inside a customer instance. The initiator for PV iSCSI attachments can be set up and run inside a cloud instance's hypervisor, and PV iSCSI attachments can be limited to running on virtual machines (VM). The initiator for NVMe attachments can be run on a smartNIC. Accordingly, NVMe attachments can provide attachments for both VM and BM networks.

FIG. 6 is a simplified control path diagram 600 showing cloud infrastructure components for attaching persistent storage, according to an embodiment, for example, using NVMe. A customer administrator 605 can submit a request for a new storage attachment at an application programming interface (API) endpoint 610. In some examples, the customer administrator 605 may be any entity that manages or otherwise administers the use of cloud instances for a customer of the cloud service. In some instances, the API endpoint 610 may be an interface where customer's (e.g., customer administrator 605) can access the cloud service resources, for example, by making requests to have operations performed by the cloud service on resources managed for the customer. The request can be forwarded to the compute control plane 615 in a compute control plane service enclave 620. In some instances, compute control plane 615 can be a series of APIs that can provision, manage, reconfigure, or terminate resources based on user requests. The request can be forwarded from compute control plane 615 to the block storage control plane 625 in the block storage control plane enclave 630. In some examples, the block storage control plane 625 can be a series of APIs that can provision, manage, reconfigure, or terminate block storage.

A request that is received at block storage control plane 625 can be forwarded to the storage cluster management plane 635. Storage cluster management plane 635 can manage the server fleets, and, for example, storage cluster management plane 635 can manage extent server fleet 640 and target server fleet 645. In some examples, storage cluster management plane 635 can configure and monitor extent server fleet 640 or target server fleet 645, and extent server fleet 640 can include servers storing striped and encrypted customer data. Volumes can be striped across multiple extent servers in extent server fleet 640. Extent servers can be a block storage data plane service that handles extent-level I/O and stores the data for replication. In response to the request, storage cluster management plane 635 can identify at least one target server 650 in the target server fleet 645 as a target server for the attachment. In some instances, target server 650 can be a server that manages the flow of customer data to and from extent server fleet 640. Target server 650 can accept I/O requests from a NVMe initiator and send the requests to extent server fleet 640. The storage cluster management plane 635 can select the target server 650 based at least in part on the load experienced by the servers in the target server fleet 645, or the expected volume for the attachment. Storage cluster management plane 635 can forward information about the new attachment to the selected target server 650 or the extent server fleet 640. The information can identify one or more target servers that are able to receive traffic from the new attachment.

The request can be forwarded from block storage control plane 625 to the block shadow service 655. The block shadow service 655 can act as an agent, and block shadow service 655 can communicate with the block smartNIC agent (BSA) 660 in smartNIC 665. In some examples, smartNIC 665 can be hardware that can connect the customer virtual network 670 to other computer networks. BSA 660 can serve as a communication link between block shadow service 655 and an NVMe agent in smartNIC 665. Communication from the block shadow service 655 can provide information about the target server and the attachment to BSA 660. A connection between the customer virtual network 670 and target server fleet 645 can be established by BSA 660. BSA 660 can expose a namespace to the host through host PCIe connection, which can be accessed by the host applications and by the customer through the customer virtual network 670. The customer virtual network 670 can be set up by the VCN, and traffic from customer virtual network 670 can reach extent servers fleet 640 via target server fleet 645 through smartNIC 665.

FIG. 7 is a diagram 700 showing a kernel architecture for implementing Internet Small Computer Systems Interface (iSCSI) and Non-Volatile Memory Express (NVMe) attachments according to an embodiment. NVMe and iSCSI are networking protocols providing block-level storage access, and both NVMe and iSCSI can be used to attach persistent storage. One difference between the two standards is that, in an iSCSI architecture, input/output (IO) requests reach a smartNIC via a host network interface card (NIC), and, in an NVMe architecture, the smartNIC is directly connected to a Peripheral Component Interconnect Express (PCIe) bus. The NVMe kernel stack can be streamlined compared to the iSCSI stack, and NVMe's simplified architecture can be achieved because the NVMe initiator can be located in the smartNIC.

In a host server 705, using either networking protocol, traffic can reach a file system 710 in the kernel 715 from an application 720 in the user space 725. The traffic can be addressed to a target 730 that can be a block storage server (e.g., target fleet 645, extent servers fleet 640, etc.). Traffic for the two standards can follow similar pathways until the traffic arrives at block 735 from file system 710.

Using iSCSI, traffic from block 735 reaches the PCIe bus 740 via SCSI 745, iSCSI initiator 750, TCP/IP 755, and the NIC driver 760. iSCSI traffic leaving PCIe bus 740 can reach the target via host NIC 765 and smartNIC 770. In some instances, PCIe bus 740 can be a serial computer expansion bus. The NVMe pathway can follow a different pathway, and NVMe traffic can reach PCIe bus 740 from block 735 via NVMe driver 775. Instead of passing through host NIC 765, NVMe traffic can travel from PCIe 740 to smartNIC 770 before reaching target 730. The NVMe initiator 780 can be located in smartNIC 770 instead of being located in kernel 715 like iSCSI initiator 750.

FIG. 8 is a Non-Volatile Memory Express (NVMe) system diagram 800 according to an embodiment. A customer, such as customer admin 605, can initiate an NVMe attachment request from the console or a public API (e.g., API end point 610). The NVMe attachment request can be forwarded from the control plane 805 (e.g. block storage control plane 625) to an agent 810 (e.g., smartNIC BSA 660) in the smartNIC processor 815. The agent 810 can perform health checks on NVMe/TCP targets 820a-c to identify healthy targets, and agent 810 can instruct the NVMe/TCP initiator 825 in the Programming Protocol-Independent Packet Processors (P4) pipeline 830 to establish a connection with a healthy NVMe/TCP target (e.g., NVMe/TCP target 820b). P4 is a domain-specific programming language that is optimized for controlling packet forwarding. NVMe/TCP initiator 825 can communicate with Storage Performance Development Kit (SPDK) reactor 835 to initiate the connection.

Once a connection is established with NVMe/TCP target 820b and the NVMe attachment is completed, virtual machine/bare metal (VM/BM) instance 840 can issue NVMe admin commands or NVMe I/O commands to the NVMe/TCP target 820b. The NVMe commands can be issued from BM/VM instance 840 to NVMe PCIe admin queue 845 or NVMe PCIe I/O queue 850 via NVMe block driver 855 and virtual function (VF) 860. In some examples, VF 860 can be a PCIe function that supports single root I/O visualization (SR-IOV). In some instances, the admin queue can be used to establish host-controller associations and the queue can support commands like Identify, Get/Set Features, etc. Agent 810 can retrieve NVMe admin commands from the NVMe PCIe admin queue 845 and forward those commands to NVMe/TCP target 820b, or the commands can be processed locally. I/O commands received from VM/BM instance 840 can be enqueued into NVMe PCIe I/O queue 850. NVMe block driver 855 can retrieve the enqueued commands from NVMe PCIe I/O queue 850 to NVMe/TCP target 820b via NVMe/TCP initiator 825.

FIG. 9 is a diagram 900 of a Non-Volatile Memory Express (NVMe)/Transmission control Protocol (TCP) target according to an embodiment. The NVMe/TCP target (e.g., NVMe/TCP target 820b) can be a Non-Uniform Memory Access (NUMA) node 905 that can include a central processing unit coupled with memory. Cores in the NUMA node 905 CPU can be assigned to one or more SPDK reactor cores such as SPDK reactor cores 910a-b (e.g., SPDK reactor 835). Accept poller 915 can accept new connections to the SPDK reactor and assign the new connections to a SPDK reactor core (e.g., SPDK reactor core 910a). Accept Poller 915 can assign new connections to an available TCP poll group 920a-b in an available SPDK reactor core 910a-b, and the new connections can be assigned using a round robin algorithm.

Subsystem controllers 925a-c can be assigned to a new connection, and, for example, subsystem controller 925a can be assigned for a connection made with TCP poll group 920a. More than one subsystem controller 925a-c can be assigned to one of the TCP poll groups 920a-b, and, for instance, subsystem controller 925a and subsystem controller 925b can be assigned to TCP poll group 920a. Block device namespaces 930a-c can be generated when a connection is made with one of the subsystem controllers 925a-c.

Threads in a NUMA node CPU can be assigned as client threads 935a-c by one of the block device namespaces. Block device namespaces 930a-c can forward a request that is received through the new connection to one of the client threads 935a-c, and client threads 935a-c can decide which extent server 940a-c should receive the data associated with the request. After completing the request, client threads 935a-c can send a response to message queue 945a-b to indicate that a request has been completed. Requests can be received at a SPDK reactor core 910a-b from the smartNIC initiator (e.g., NVMe/TCP initiator 825, NVMe initiator 780, etc.) or a different initiator (e.g., iSCSI initiator 750). Responses can be sent from one of the SPDK reactor cores 910a-b to the smartNIC initiator or a different initiator.

FIG. 10 is a simplified diagram 1000 of a smart network interface card (smartNIC) with Non-Volatile Memory Express (NVMe) according to an embodiment. Requests can be received at smartNIC 1005 from the block storage shadow service 1010 (e.g., block storage shadow service 655) in the control plane (e.g., block storage control plane 625, control plane 805, etc.). The requests can be received at the block smartNIC agent (BSA) 1015 (e.g., BSA 660) running on the smartNIC central processing unit (CPU) 1020. BSA 1015 can serve a number of functions including performing health checks, ensuring that targets are available, or performing telemetry. BSA 1015 forwards instructions or requests to the host 1025, or other smartNIC components, via NVMe Agent 1030. Requests or instructions can be sent from NVMe Agent 1030 to the NVMe driver 1035 via a PCIe physical function or virtual function (PF/VF) 1040 (e.g., VF 860).

The NVMe agent 1030 can establish a new I/O connection in response to a request from BSA 1015 using the vector packet processing/dataplane development kit (VPP/DPDK) module 1045. The VPP/DPDK module can use a framework, such as VPP with the DPDK plugin, to process and route network packets. Upon receiving a request from NVMe agent 1030, VPP/DPDK 1045 can send a request to the P4 pipeline 1050 (e.g., P4 pipeline 830) via the Ethernet (ETH) P4 module 1055 running on the P4 match protection unit (MPU) 1060. P4 pipeline 1050 can establish an I/O connection with SPDK NVMe/TCP targets 1065 (e.g., target 730, target fleet 645, NVMe/TCP target 820a-c, etc.). Establishing a connection can include sending instructions to NVMe driver 1035 or SPDK NVMe/TCP targets 1065

The I/O communication can be offloaded to a fast path I/O pipeline after an I/O connection is established with an SPDK NVMe/TCP target 1065. The I/O fast path traffic can travel along the fast path pipeline from the I/O submission queue/completion queue (SQ/CQ) 1070 in host 1025 to P4 MPUs 1060 via PCIe PF/VF 1040. I/O traffic can be received in P4 MPUs 1060 at NVMe P4 1075 and forwarded to the SPDK NVMe/TCP targets 1065 via TCP P4 1080 and P4 pipeline 1050. Traffic in I/O SQ/CQ 1070 can start from the submission queue and end at the completion queue when I/O completes. If traffic along the fast path pipeline fails, NVMe P4 1075 or TCP P4 1080 can inform NVMe Agent 1030 of the failure. NVMe agent 1030 can be configured so that NVMe agent can create a new I/O connection in response to the failure and offload the new connection to the fast path pipeline. XTS engine 1085 is an encryption engine that can encrypt user data using the xor-encrypt-xor (XEX)-based tweaked-codebook mode with ciphertext stealing (XTS) block cypher, and hash engine 1090 can use cryptographic hash functions to verify data integrity.

FIG. 11 is a diagram 1100 showing multipath handling in a smart network interface card (smartNIC) according to an embodiment. An application 1105 can run in a virtual machine (VM) 1110 managed by a hypervisor 1115. Application 1105 can be similar to application 720, and VM 1110 can be a bare metal machine (e.g., BM/VM instance 840). Two namespaces devices, namespace 1 device 1120 and namespace 7 device 1125, can be associated with Application 1105. A namespace can be a NVM storage that is formatted for block access. A namespace can be analogous to a logical unit in SCSI, and a block storage volume can be a single namespace. Traffic between namespace 1 1120 or namespace 7 1125 and the NVMe/TCP target servers 1130a-i (e.g., target server 650) can be received via the virtual function input/output queue (VFIO) 1135 in the kernel 1140. The virtual function (VF) 1145 can be connected to VFIO 1135 via the VFIO peripheral component interconnect (PCI) 1150. VF 1145 can be a virtual function or a physical function.

The NVMe/PCIe controller 1155 can route traffic from the namespace devices to NVMe namespaces. For instance, traffic can be routed between namespace device 1 1120 and NVMe namespace 1 1160, and traffic can be routed between namespace device 7 1125 and NVMe namespace 7 1165. The namespace devices can be associated with one or more path groups 1170a-d located in the P4 pipeline 1175 (e.g., P4 pipeline 1050, P4 MPUs 1060, etc.) in smartNIC 1180 (e.g., smartNIC 665, smartNIC 770, smartNIC 1005, etc.). For instance, NVMe namespace 1 1160 can route traffic to path groups 1170a-c, and NVMe namespace 7 1165 can route traffic to path group 1170d.

Path groups can contain an active path 1180a-d and one or more passive paths 1185a-h. Active paths 1180a-d or passive paths 1185a-h can be associated with a NVMe/TCP target server 1130a-i. Traffic between a NVMe/TCP target server 1130a-i and namespace device 1 1120 or namespace device 7 1125 can be routed via active paths 1180a-d. NVMe/TCP target servers 1130a-i can route traffic to and from extent servers (e.g., extent servers fleet 640, extent servers 940a-c, etc.).

Traffic can be routed via a passive path 1185a-h if an active path 1180a-d fails. In response to a failure, passive path 1185a-h can login to an extent server via NVMe/TCP target servers 1130a-h. The extent server can change a token from the token for an active path 1180a-d to a token for a passive path 1185a-h. The extent server can use the token to determine whether to accept traffic from a path (e.g., active paths 1180a-d or passive paths 1185a-h).

FIG. 12 shows a diagram of an architecture 1200 for performing encryption/decryption with a smart network interface card (smartNIC) according to an embodiment. The architecture 1200 can provide a unified means for encrypting/decrypting both VM and BM traffic. NVMe driver 1205a can run in the kernel 1210a of a bare metal (BM) machine 1215 (e.g., BM/VM instance 840, etc.). Traffic can be sent from NVMe driver to SPDK NVMe/TCP targets 1220 via smartNIC 1225a. The BM traffic can be received via a physical function (PF) 1230 (e.g., PCIe PF/VF 1040, etc.) at the NVMe PCI controller 1235a (e.g., NVMe/PCIe controller 1155, NVMe P4 1075, etc.) in the P4 pipeline 1240a (e.g., P4 MPUs 1060, P4 pipeline 1050, etc.).

Outgoing BM traffic traveling from NVMe driver 1205a to SPDK NVMe/TCP targets 1220 can be encrypted by the encryption module 1245a in smartNIC 1225a, and incoming BM traffic can be decrypted by the encryption module 1245a. Encryption module 1245a can encrypt or decrypt traffic using an encryption algorithm such as Advanced Encryption Standard (AES). The encrypted BM traffic can be sent to SPDK NVMe/TCP targets 1220 via the NVMe/TCP initiator 1250a (e.g., NVMe initiator 780, NVMe/TCP initiator 825, etc.). Incoming encrypted BM traffic from SPDK NVMe/TCP targets 1220 can be received at NVMe/TCP initiator 1250a before being forwarded along the pathway to NVMe driver 1205a. Incoming encrypted BM traffic can be decrypted by the encryption module 1245a.

Outgoing VM traffic can be sent from NVMe driver 1205b in the virtual machine (VM) 1255 (e.g., BM/VM instance 840, VM 1110, etc.) to the virtual function input/output (VFIO) 1260 in kernel 1210b via a virtual function (VF) 1260 (e.g., VF 860, VF 1145, etc.). The outgoing VM traffic can be forwarded to NVME PCI controller 1235b in P4 pipeline 1240b. The outgoing VM traffic can be forwarded from smartNIC 1225b to SPDK NVMe/TCP targets 1220 via encryption module 1245b and NVMe/TCP initiator 1250b. Incoming VM traffic from SPDK NVMe/TCP targets 1220 can be received at NVMe/TCP initiator 1250b before the incoming traffic is forwarded along the pathway to NVMe driver 1205b. Incoming encrypted VM traffic can be decrypted by the encryption module 1245a. FIG. 13 is a block diagram 1300 illustrating an example pattern of an IaaS architecture, according to at least one embodiment. Service operators 1302 can be communicatively coupled to a secure host tenancy 1304 that can include a virtual cloud network (VCN) 1306 and a secure host subnet 1308. In some examples, the service operators 1302 may be using one or more client computing devices, which may be portable handheld devices (e.g., an iPhone®, cellular telephone, an iPad®, computing tablet, a personal digital assistant (PDA)) or wearable devices (e.g., a Google Glass® head mounted display), running software such as Microsoft Windows Mobile®, and/or a variety of mobile operating systems such as iOS, Windows Phone, Android, BlackBerry 8, Palm OS, and the like, and being Internet, e-mail, short message service (SMS), Blackberry®, or other communication protocol enabled. Alternatively, the client computing devices can be general purpose personal computers including, by way of example, personal computers and/or laptop computers running various versions of Microsoft Windows®, Apple Macintosh®, and/or Linux operating systems. The client computing devices can be workstation computers running any of a variety of commercially-available UNIX® or UNIX-like operating systems, including without limitation the variety of GNU/Linux operating systems, such as for example, Google Chrome OS. Alternatively, or in addition, client computing devices may be any other electronic device, such as a thin-client computer, an Internet-enabled gaming system (e.g., a Microsoft Xbox gaming console with or without a Kinect® gesture input device), and/or a personal messaging device, capable of communicating over a network that can access the VCN 1306 and/or the Internet.

The VCN 1306 can include a local peering gateway (LPG) 1310 that can be communicatively coupled to a secure shell (SSH) VCN 1312 via an LPG 1310 contained in the SSH VCN 1312. The SSH VCN 1312 can include an SSH subnet 1314, and the SSH VCN 1312 can be communicatively coupled to a control plane VCN 1316 via the LPG 1310 contained in the control plane VCN 1316. Also, the SSH VCN 1312 can be communicatively coupled to a data plane VCN 1318 via an LPG 1310. The control plane VCN 1316 and the data plane VCN 1318 can be contained in a service tenancy 1319 that can be owned and/or operated by the IaaS provider.

The control plane VCN 1316 can include a control plane demilitarized zone (DMZ) tier 1320 that acts as a perimeter network (e.g., portions of a corporate network between the corporate intranet and external networks). The DMZ-based servers may have restricted responsibilities and help keep breaches contained. Additionally, the DMZ tier 1320 can include one or more load balancer (LB) subnet(s) 1322, a control plane app tier 1324 that can include app subnet(s) 1326, a control plane data tier 1328 that can include database (DB) subnet(s) 1330 (e.g., frontend DB subnet(s) and/or backend DB subnet(s)). The LB subnet(s) 1322 contained in the control plane DMZ tier 1320 can be communicatively coupled to the app subnet(s) 1326 contained in the control plane app tier 1324 and an Internet gateway 1334 that can be contained in the control plane VCN 1316, and the app subnet(s) 1326 can be communicatively coupled to the DB subnet(s) 1330 contained in the control plane data tier 1328 and a service gateway 1336 and a network address translation (NAT) gateway 1338. The control plane VCN 1316 can include the service gateway 1336 and the NAT gateway 1338.

The control plane VCN 1316 can include a data plane mirror app tier 1340 that can include app subnet(s) 1326. The app subnet(s) 1326 contained in the data plane mirror app tier 1340 can include a virtual network interface controller (VNIC) 1342 that can execute a compute instance 1344. The compute instance 1344 can communicatively couple the app subnet(s) 1326 of the data plane mirror app tier 1340 to app subnet(s) 1326 that can be contained in a data plane app tier 1346.

The data plane VCN 1318 can include the data plane app tier 1346, a data plane DMZ tier 1348, and a data plane data tier 1350. The data plane DMZ tier 1348 can include LB subnet(s) 1322 that can be communicatively coupled to the app subnet(s) 1326 of the data plane app tier 1346 and the Internet gateway 1334 of the data plane VCN 1318. The app subnet(s) 1326 can be communicatively coupled to the service gateway 1336 of the data plane VCN 1318 and the NAT gateway 1338 of the data plane VCN 1318. The data plane data tier 1350 can also include the DB subnet(s) 1330 that can be communicatively coupled to the app subnet(s) 1326 of the data plane app tier 1346.

The Internet gateway 1334 of the control plane VCN 1316 and of the data plane VCN 1318 can be communicatively coupled to a metadata management service 1352 that can be communicatively coupled to public Internet 1354. Public Internet 1354 can be communicatively coupled to the NAT gateway 1338 of the control plane VCN 1316 and of the data plane VCN 1318. The service gateway 1336 of the control plane VCN 1316 and of the data plane VCN 1318 can be communicatively couple to cloud services 1356.

In some examples, the service gateway 1336 of the control plane VCN 1316 or of the data plane VCN 1318 can make application programming interface (API) calls to cloud services 1356 without going through public Internet 1354. The API calls to cloud services 1356 from the service gateway 1336 can be one-way: the service gateway 1336 can make API calls to cloud services 1356, and cloud services 1356 can send requested data to the service gateway 1336. But, cloud services 1356 may not initiate API calls to the service gateway 1336.

In some examples, the secure host tenancy 1304 can be directly connected to the service tenancy 1319, which may be otherwise isolated. The secure host subnet 1308 can communicate with the SSH subnet 1314 through an LPG 1310 that may enable two-way communication over an otherwise isolated system. Connecting the secure host subnet 1308 to the SSH subnet 1314 may give the secure host subnet 1308 access to other entities within the service tenancy 1319.

The control plane VCN 1316 may allow users of the service tenancy 1319 to set up or otherwise provision desired resources. Desired resources provisioned in the control plane VCN 1316 may be deployed or otherwise used in the data plane VCN 1318. In some examples, the control plane VCN 1316 can be isolated from the data plane VCN 1318, and the data plane mirror app tier 1340 of the control plane VCN 1316 can communicate with the data plane app tier 1346 of the data plane VCN 1318 via VNICs 1342 that can be contained in the data plane mirror app tier 1340 and the data plane app tier 1346.

In some examples, users of the system, or customers, can make requests, for example create, read, update, or delete (CRUD) operations, through public Internet 1354 that can communicate the requests to the metadata management service 1352. The metadata management service 1352 can communicate the request to the control plane VCN 1316 through the Internet gateway 1334. The request can be received by the LB subnet(s) 1322 contained in the control plane DMZ tier 1320. The LB subnet(s) 1322 may determine that the request is valid, and in response to this determination, the LB subnet(s) 1322 can transmit the request to app subnet(s) 1326 contained in the control plane app tier 1324. If the request is validated and requires a call to public Internet 1354, the call to public Internet 1354 may be transmitted to the NAT gateway 1338 that can make the call to public Internet 1354. Metadata that may be desired to be stored by the request can be stored in the DB subnet(s) 1330.

In some examples, the data plane mirror app tier 1340 can facilitate direct communication between the control plane VCN 1316 and the data plane VCN 1318. For example, changes, updates, or other suitable modifications to configuration may be desired to be applied to the resources contained in the data plane VCN 1318. Via a VNIC 1342, the control plane VCN 1316 can directly communicate with, and can thereby execute the changes, updates, or other suitable modifications to configuration to, resources contained in the data plane VCN 1318.

In some embodiments, the control plane VCN 1316 and the data plane VCN 1318 can be contained in the service tenancy 1319. In this case, the user, or the customer, of the system may not own or operate either the control plane VCN 1316 or the data plane VCN 1318. Instead, the IaaS provider may own or operate the control plane VCN 1316 and the data plane VCN 1318, both of which may be contained in the service tenancy 1319. This embodiment can enable isolation of networks that may prevent users or customers from interacting with other users', or other customers', resources. Also, this embodiment may allow users or customers of the system to store databases privately without needing to rely on public Internet 1354, which may not have a desired level of threat prevention, for storage.

In other embodiments, the LB subnet(s) 1322 contained in the control plane VCN 1316 can be configured to receive a signal from the service gateway 1336. In this embodiment, the control plane VCN 1316 and the data plane VCN 1318 may be configured to be called by a customer of the IaaS provider without calling public Internet 1354. Customers of the IaaS provider may desire this embodiment since database(s) that the customers use may be controlled by the IaaS provider and may be stored on the service tenancy 1319, which may be isolated from public Internet 1354.

FIG. 14 is a block diagram 1400 illustrating another example pattern of an IaaS architecture, according to at least one embodiment. Service operators 1402 (e.g., service operators 1302 of FIG. 13) can be communicatively coupled to a secure host tenancy 1404 (e.g., the secure host tenancy 1304 of FIG. 13) that can include a virtual cloud network (VCN) 1406 (e.g., the VCN 1306 of FIG. 13) and a secure host subnet 1408 (e.g., the secure host subnet 1308 of FIG. 13). The VCN 1406 can include a local peering gateway (LPG) 1410 (e.g., the LPG 1310 of FIG. 13) that can be communicatively coupled to a secure shell (SSH) VCN 1412 (e.g., the SSH VCN 1312 of FIG. 13) via an LPG 1310 contained in the SSH VCN 1412. The SSH VCN 1412 can include an SSH subnet 1414 (e.g., the SSH subnet 1314 of FIG. 13), and the SSH VCN 1412 can be communicatively coupled to a control plane VCN 1416 (e.g., the control plane VCN 1316 of FIG. 13) via an LPG 1410 contained in the control plane VCN 1416. The control plane VCN 1416 can be contained in a service tenancy 1419 (e.g., the service tenancy 1319 of FIG. 13), and the data plane VCN 1418 (e.g., the data plane VCN 1318 of FIG. 13) can be contained in a customer tenancy 1421 that may be owned or operated by users, or customers, of the system.

The control plane VCN 1416 can include a control plane DMZ tier 1420 (e.g., the control plane DMZ tier 1320 of FIG. 13) that can include LB subnet(s) 1422 (e.g., LB subnet(s) 1322 of FIG. 13), a control plane app tier 1424 (e.g., the control plane app tier 1324 of FIG. 13) that can include app subnet(s) 1426 (e.g., app subnet(s) 1326 of FIG. 13), a control plane data tier 1428 (e.g., the control plane data tier 1328 of FIG. 13) that can include database (DB) subnet(s) 1430 (e.g., similar to DB subnet(s) 1330 of FIG. 13). The LB subnet(s) 1422 contained in the control plane DMZ tier 1420 can be communicatively coupled to the app subnet(s) 1426 contained in the control plane app tier 1424 and an Internet gateway 1434 (e.g., the Internet gateway 1334 of FIG. 13) that can be contained in the control plane VCN 1416, and the app subnet(s) 1426 can be communicatively coupled to the DB subnet(s) 1430 contained in the control plane data tier 1428 and a service gateway 1436 (e.g., the service gateway 1336 of FIG. 13) and a network address translation (NAT) gateway 1438 (e.g., the NAT gateway 1338 of FIG. 13). The control plane VCN 1416 can include the service gateway 1436 and the NAT gateway 1438.

The control plane VCN 1416 can include a data plane mirror app tier 1440 (e.g., the data plane mirror app tier 1340 of FIG. 13) that can include app subnet(s) 1426. The app subnet(s) 1426 contained in the data plane mirror app tier 1440 can include a virtual network interface controller (VNIC) 1442 (e.g., the VNIC of 1342) that can execute a compute instance 1444 (e.g., similar to the compute instance 1344 of FIG. 13). The compute instance 1444 can facilitate communication between the app subnet(s) 1426 of the data plane mirror app tier 1440 and the app subnet(s) 1426 that can be contained in a data plane app tier 1446 (e.g., the data plane app tier 1346 of FIG. 13) via the VNIC 1442 contained in the data plane mirror app tier 1440 and the VNIC 1442 contained in the data plane app tier 1446.

The Internet gateway 1434 contained in the control plane VCN 1416 can be communicatively coupled to a metadata management service 1452 (e.g., the metadata management service 1352 of FIG. 13) that can be communicatively coupled to public Internet 1454 (e.g., public Internet 1354 of FIG. 13). Public Internet 1454 can be communicatively coupled to the NAT gateway 1438 contained in the control plane VCN 1416. The service gateway 1436 contained in the control plane VCN 1416 can be communicatively couple to cloud services 1456 (e.g., cloud services 1356 of FIG. 13).

In some examples, the data plane VCN 1418 can be contained in the customer tenancy 1421. In this case, the IaaS provider may provide the control plane VCN 1416 for each customer, and the IaaS provider may, for each customer, set up a unique compute instance 1444 that is contained in the service tenancy 1419. Each compute instance 1444 may allow communication between the control plane VCN 1416, contained in the service tenancy 1419, and the data plane VCN 1418 that is contained in the customer tenancy 1421. The compute instance 1444 may allow resources, that are provisioned in the control plane VCN 1416 that is contained in the service tenancy 1419, to be deployed or otherwise used in the data plane VCN 1418 that is contained in the customer tenancy 1421.

In other examples, the customer of the IaaS provider may have databases that live in the customer tenancy 1421. In this example, the control plane VCN 1416 can include the data plane mirror app tier 1440 that can include app subnet(s) 1426. The data plane mirror app tier 1440 can reside in the data plane VCN 1418, but the data plane mirror app tier 1440 may not live in the data plane VCN 1418. That is, the data plane mirror app tier 1440 may have access to the customer tenancy 1421, but the data plane mirror app tier 1440 may not exist in the data plane VCN 1418 or be owned or operated by the customer of the IaaS provider. The data plane mirror app tier 1440 may be configured to make calls to the data plane VCN 1418 but may not be configured to make calls to any entity contained in the control plane VCN 1416. The customer may desire to deploy or otherwise use resources in the data plane VCN 1418 that are provisioned in the control plane VCN 1416, and the data plane mirror app tier 1440 can facilitate the desired deployment, or other usage of resources, of the customer.

In some embodiments, the customer of the IaaS provider can apply filters to the data plane VCN 1418. In this embodiment, the customer can determine what the data plane VCN 1418 can access, and the customer may restrict access to public Internet 1454 from the data plane VCN 1418. The IaaS provider may not be able to apply filters or otherwise control access of the data plane VCN 1418 to any outside networks or databases. Applying filters and controls by the customer onto the data plane VCN 1418, contained in the customer tenancy 1421, can help isolate the data plane VCN 1418 from other customers and from public Internet 1454.

In some embodiments, cloud services 1456 can be called by the service gateway 1436 to access services that may not exist on public Internet 1454, on the control plane VCN 1416, or on the data plane VCN 1418. The connection between cloud services 1456 and the control plane VCN 1416 or the data plane VCN 1418 may not be live or continuous. Cloud services 1456 may exist on a different network owned or operated by the IaaS provider. Cloud services 1456 may be configured to receive calls from the service gateway 1436 and may be configured to not receive calls from public Internet 1454. Some cloud services 1456 may be isolated from other cloud services 1456, and the control plane VCN 1416 may be isolated from cloud services 1456 that may not be in the same region as the control plane VCN 1416. For example, the control plane VCN 1416 may be located in “Region 1,” and cloud service “Deployment 13,” may be located in Region 1 and in “Region 2.” If a call to Deployment 13 is made by the service gateway 1436 contained in the control plane VCN 1416 located in Region 1, the call may be transmitted to Deployment 13 in Region 1. In this example, the control plane VCN 1416, or Deployment 13 in Region 1, may not be communicatively coupled to, or otherwise in communication with, Deployment 13 in Region 2.

FIG. 15 is a block diagram 1500 illustrating another example pattern of an IaaS architecture, according to at least one embodiment. Service operators 1502 (e.g., service operators 1302 of FIG. 13) can be communicatively coupled to a secure host tenancy 1504 (e.g., the secure host tenancy 1304 of FIG. 13) that can include a virtual cloud network (VCN) 1506 (e.g., the VCN 1306 of FIG. 13) and a secure host subnet 1508 (e.g., the secure host subnet 1308 of FIG. 13). The VCN 1506 can include an LPG 1510 (e.g., the LPG 1310 of FIG. 13) that can be communicatively coupled to an SSH VCN 1512 (e.g., the SSH VCN 1312 of FIG. 13) via an LPG 1510 contained in the SSH VCN 1512. The SSH VCN 1512 can include an SSH subnet 1514 (e.g., the SSH subnet 1314 of FIG. 13), and the SSH VCN 1512 can be communicatively coupled to a control plane VCN 1516 (e.g., the control plane VCN 1316 of FIG. 13) via an LPG 1510 contained in the control plane VCN 1516 and to a data plane VCN 1518 (e.g., the data plane 1318 of FIG. 13) via an LPG 1510 contained in the data plane VCN 1518. The control plane VCN 1516 and the data plane VCN 1518 can be contained in a service tenancy 1519 (e.g., the service tenancy 1319 of FIG. 13).

The control plane VCN 1516 can include a control plane DMZ tier 1520 (e.g., the control plane DMZ tier 1320 of FIG. 13) that can include load balancer (LB) subnet(s) 1522 (e.g., LB subnet(s) 1322 of FIG. 13), a control plane app tier 1524 (e.g., the control plane app tier 1324 of FIG. 13) that can include app subnet(s) 1526 (e.g., similar to app subnet(s) 1326 of FIG. 13), a control plane data tier 1528 (e.g., the control plane data tier 1328 of FIG. 13) that can include DB subnet(s) 1530. The LB subnet(s) 1522 contained in the control plane DMZ tier 1520 can be communicatively coupled to the app subnet(s) 1526 contained in the control plane app tier 1524 and to an Internet gateway 1534 (e.g., the Internet gateway 1334 of FIG. 13) that can be contained in the control plane VCN 1516, and the app subnet(s) 1526 can be communicatively coupled to the DB subnet(s) 1530 contained in the control plane data tier 1528 and to a service gateway 1536 (e.g., the service gateway of FIG. 13) and a network address translation (NAT) gateway 1538 (e.g., the NAT gateway 1338 of FIG. 13). The control plane VCN 1516 can include the service gateway 1536 and the NAT gateway 1538.

The data plane VCN 1518 can include a data plane app tier 1546 (e.g., the data plane app tier 1346 of FIG. 13), a data plane DMZ tier 1548 (e.g., the data plane DMZ tier 1348 of FIG. 13), and a data plane data tier 1550 (e.g., the data plane data tier 1350 of FIG. 13). The data plane DMZ tier 1548 can include LB subnet(s) 1522 that can be communicatively coupled to trusted app subnet(s) 1560 and untrusted app subnet(s) 1562 of the data plane app tier 1546 and the Internet gateway 1534 contained in the data plane VCN 1518. The trusted app subnet(s) 1560 can be communicatively coupled to the service gateway 1536 contained in the data plane VCN 1518, the NAT gateway 1538 contained in the data plane VCN 1518, and DB subnet(s) 1530 contained in the data plane data tier 1550. The untrusted app subnet(s) 1562 can be communicatively coupled to the service gateway 1536 contained in the data plane VCN 1518 and DB subnet(s) 1530 contained in the data plane data tier 1550. The data plane data tier 1550 can include DB subnet(s) 1530 that can be communicatively coupled to the service gateway 1536 contained in the data plane VCN 1518.

The untrusted app subnet(s) 1562 can include one or more primary VNICs 1564(1)-(N) that can be communicatively coupled to tenant virtual machines (VMs) 1566(1)-(N). Each tenant VM 1566(1)-(N) can be communicatively coupled to a respective app subnet 1567(1)-(N) that can be contained in respective container egress VCNs 1568(1)-(N) that can be contained in respective customer tenancies 1570(1)-(N). Respective secondary VNICs 1572(1)-(N) can facilitate communication between the untrusted app subnet(s) 1562 contained in the data plane VCN 1518 and the app subnet contained in the container egress VCNs 1568(1)-(N). Each container egress VCNs 1568(1)-(N) can include a NAT gateway 1538 that can be communicatively coupled to public Internet 1554 (e.g., public Internet 1354 of FIG. 13).

The Internet gateway 1534 contained in the control plane VCN 1516 and contained in the data plane VCN 1518 can be communicatively coupled to a metadata management service 1552 (e.g., the metadata management system 1352 of FIG. 13) that can be communicatively coupled to public Internet 1554. Public Internet 1554 can be communicatively coupled to the NAT gateway 1538 contained in the control plane VCN 1516 and contained in the data plane VCN 1518. The service gateway 1536 contained in the control plane VCN 1516 and contained in the data plane VCN 1518 can be communicatively couple to cloud services 1556.

In some embodiments, the data plane VCN 1518 can be integrated with customer tenancies 1570. This integration can be useful or desirable for customers of the IaaS provider in some cases such as a case that may desire support when executing code. The customer may provide code to run that may be destructive, may communicate with other customer resources, or may otherwise cause undesirable effects. In response to this, the IaaS provider may determine whether to run code given to the IaaS provider by the customer.

In some examples, the customer of the IaaS provider may grant temporary network access to the IaaS provider and request a function to be attached to the data plane app tier 1546. Code to run the function may be executed in the VMs 1566(1)-(N), and the code may not be configured to run anywhere else on the data plane VCN 1518. Each VM 1566(1)-(N) may be connected to one customer tenancy 1570. Respective containers 1571(1)-(N) contained in the VMs 1566(1)-(N) may be configured to run the code. In this case, there can be a dual isolation (e.g., the containers 1571(1)-(N) running code, where the containers 1571(1)-(N) may be contained in at least the VM 1566(1)-(N) that are contained in the untrusted app subnet(s) 1562), which may help prevent incorrect or otherwise undesirable code from damaging the network of the IaaS provider or from damaging a network of a different customer. The containers 1571(1)-(N) may be communicatively coupled to the customer tenancy 1570 and may be configured to transmit or receive data from the customer tenancy 1570. The containers 1571(1)-(N) may not be configured to transmit or receive data from any other entity in the data plane VCN 1518. Upon completion of running the code, the IaaS provider may kill or otherwise dispose of the containers 1571(1)-(N).

In some embodiments, the trusted app subnet(s) 1560 may run code that may be owned or operated by the IaaS provider. In this embodiment, the trusted app subnet(s) 1560 may be communicatively coupled to the DB subnet(s) 1530 and be configured to execute CRUD operations in the DB subnet(s) 1530. The untrusted app subnet(s) 1562 may be communicatively coupled to the DB subnet(s) 1530, but in this embodiment, the untrusted app subnet(s) may be configured to execute read operations in the DB subnet(s) 1530. The containers 1571(1)-(N) that can be contained in the VM 1566(1)-(N) of each customer and that may run code from the customer may not be communicatively coupled with the DB subnet(s) 1530.

In other embodiments, the control plane VCN 1516 and the data plane VCN 1518 may not be directly communicatively coupled. In this embodiment, there may be no direct communication between the control plane VCN 1516 and the data plane VCN 1518. However, communication can occur indirectly through at least one method. An LPG 1510 may be established by the IaaS provider that can facilitate communication between the control plane VCN 1516 and the data plane VCN 1518. In another example, the control plane VCN 1516 or the data plane VCN 1518 can make a call to cloud services 1556 via the service gateway 1536. For example, a call to cloud services 1556 from the control plane VCN 1516 can include a request for a service that can communicate with the data plane VCN 1518.

FIG. 16 is a block diagram 1600 illustrating another example pattern of an IaaS architecture, according to at least one embodiment. Service operators 1602 (e.g., service operators 1302 of FIG. 13) can be communicatively coupled to a secure host tenancy 1604 (e.g., the secure host tenancy 1304 of FIG. 13) that can include a virtual cloud network (VCN) 1606 (e.g., the VCN 1306 of FIG. 13) and a secure host subnet 1608 (e.g., the secure host subnet 1308 of FIG. 13). The VCN 1606 can include an LPG 1610 (e.g., the LPG 1310 of FIG. 13) that can be communicatively coupled to an SSH VCN 1612 (e.g., the SSH VCN 1312 of FIG. 13) via an LPG 1610 contained in the SSH VCN 1612. The SSH VCN 1612 can include an SSH subnet 1614 (e.g., the SSH subnet 1314 of FIG. 13), and the SSH VCN 1612 can be communicatively coupled to a control plane VCN 1616 (e.g., the control plane VCN 1316 of FIG. 13) via an LPG 1610 contained in the control plane VCN 1616 and to a data plane VCN 1618 (e.g., the data plane 1318 of FIG. 13) via an LPG 1610 contained in the data plane VCN 1618. The control plane VCN 1616 and the data plane VCN 1618 can be contained in a service tenancy 1619 (e.g., the service tenancy 1319 of FIG. 13).

The control plane VCN 1616 can include a control plane DMZ tier 1620 (e.g., the control plane DMZ tier 1320 of FIG. 13) that can include LB subnet(s) 1622 (e.g., LB subnet(s) 1322 of FIG. 13), a control plane app tier 1624 (e.g., the control plane app tier 1324 of FIG. 13) that can include app subnet(s) 1626 (e.g., app subnet(s) 1326 of FIG. 13), a control plane data tier 1628 (e.g., the control plane data tier 1328 of FIG. 13) that can include DB subnet(s) 1630 (e.g., DB subnet(s) 1530 of FIG. 15). The LB subnet(s) 1622 contained in the control plane DMZ tier 1620 can be communicatively coupled to the app subnet(s) 1626 contained in the control plane app tier 1624 and to an Internet gateway 1634 (e.g., the Internet gateway 1334 of FIG. 13) that can be contained in the control plane VCN 1616, and the app subnet(s) 1626 can be communicatively coupled to the DB subnet(s) 1630 contained in the control plane data tier 1628 and to a service gateway 1636 (e.g., the service gateway of FIG. 13) and a network address translation (NAT) gateway 1638 (e.g., the NAT gateway 1338 of FIG. 13). The control plane VCN 1616 can include the service gateway 1636 and the NAT gateway 1638.

The data plane VCN 1618 can include a data plane app tier 1646 (e.g., the data plane app tier 1346 of FIG. 13), a data plane DMZ tier 1648 (e.g., the data plane DMZ tier 1348 of FIG. 13), and a data plane data tier 1650 (e.g., the data plane data tier 1350 of FIG. 13). The data plane DMZ tier 1648 can include LB subnet(s) 1622 that can be communicatively coupled to trusted app subnet(s) 1660 (e.g., trusted app subnet(s) 1560 of FIG. 15) and untrusted app subnet(s) 1662 (e.g., untrusted app subnet(s) 1562 of FIG. 15) of the data plane app tier 1646 and the Internet gateway 1634 contained in the data plane VCN 1618. The trusted app subnet(s) 1660 can be communicatively coupled to the service gateway 1636 contained in the data plane VCN 1618, the NAT gateway 1638 contained in the data plane VCN 1618, and DB subnet(s) 1630 contained in the data plane data tier 1650. The untrusted app subnet(s) 1662 can be communicatively coupled to the service gateway 1636 contained in the data plane VCN 1618 and DB subnet(s) 1630 contained in the data plane data tier 1650. The data plane data tier 1650 can include DB subnet(s) 1630 that can be communicatively coupled to the service gateway 1636 contained in the data plane VCN 1618.

The untrusted app subnet(s) 1662 can include primary VNICs 1664(1)-(N) that can be communicatively coupled to tenant virtual machines (VMs) 1666(1)-(N) residing within the untrusted app subnet(s) 1662. Each tenant VM 1666(1)-(N) can run code in a respective container 1667(1)-(N), and be communicatively coupled to an app subnet 1626 that can be contained in a data plane app tier 1646 that can be contained in a container egress VCN 1668. Respective secondary VNICs 1672(1)-(N) can facilitate communication between the untrusted app subnet(s) 1662 contained in the data plane VCN 1618 and the app subnet contained in the container egress VCN 1668. The container egress VCN can include a NAT gateway 1638 that can be communicatively coupled to public Internet 1654 (e.g., public Internet 1354 of FIG. 13).

The Internet gateway 1634 contained in the control plane VCN 1616 and contained in the data plane VCN 1618 can be communicatively coupled to a metadata management service 1652 (e.g., the metadata management system 1352 of FIG. 13) that can be communicatively coupled to public Internet 1654. Public Internet 1654 can be communicatively coupled to the NAT gateway 1638 contained in the control plane VCN 1616 and contained in the data plane VCN 1618. The service gateway 1636 contained in the control plane VCN 1616 and contained in the data plane VCN 1618 can be communicatively couple to cloud services 1656.

In some examples, the pattern illustrated by the architecture of block diagram 1600 of FIG. 16 may be considered an exception to the pattern illustrated by the architecture of block diagram 1500 of FIG. 15 and may be desirable for a customer of the IaaS provider if the IaaS provider cannot directly communicate with the customer (e.g., a disconnected region). The respective containers 1667(1)-(N) that are contained in the VMs 1666(1)-(N) for each customer can be accessed in real-time by the customer. The containers 1667(1)-(N) may be configured to make calls to respective secondary VNICs 1672(1)-(N) contained in app subnet(s) 1626 of the data plane app tier 1646 that can be contained in the container egress VCN 1668. The secondary VNICs 1672(1)-(N) can transmit the calls to the NAT gateway 1638 that may transmit the calls to public Internet 1654. In this example, the containers 1667(1)-(N) that can be accessed in real-time by the customer can be isolated from the control plane VCN 1616 and can be isolated from other entities contained in the data plane VCN 1618. The containers 1667(1)-(N) may also be isolated from resources from other customers.

In other examples, the customer can use the containers 1667(1)-(N) to call cloud services 1656. In this example, the customer may run code in the containers 1667(1)-(N) that requests a service from cloud services 1656. The containers 1667(1)-(N) can transmit this request to the secondary VNICs 1672(1)-(N) that can transmit the request to the NAT gateway that can transmit the request to public Internet 1654. Public Internet 1654 can transmit the request to LB subnet(s) 1622 contained in the control plane VCN 1616 via the Internet gateway 1634. In response to determining the request is valid, the LB subnet(s) can transmit the request to app subnet(s) 1626 that can transmit the request to cloud services 1656 via the service gateway 1636.

It should be appreciated that IaaS architectures 1300, 1400, 1500, 1600 depicted in the figures may have other components than those depicted. Further, the embodiments shown in the figures are only some examples of a cloud infrastructure system that may incorporate an embodiment of the disclosure. In some other embodiments, the IaaS systems may have more or fewer components than shown in the figures, may combine two or more components, or may have a different configuration or arrangement of components.

In certain embodiments, the IaaS systems described herein may include a suite of applications, middleware, and database service offerings that are delivered to a customer in a self-service, subscription-based, elastically scalable, reliable, highly available, and secure manner. An example of such an IaaS system is the Oracle Cloud Infrastructure (OCI) provided by the present assignee.

FIG. 17 illustrates an example computer system 1700, in which various embodiments may be implemented. The system 1700 may be used to implement any of the computer systems described above. As shown in the figure, computer system 1700 includes a processing unit 1704 that communicates with a number of peripheral subsystems via a bus subsystem 1702. These peripheral subsystems may include a processing acceleration unit 1706, an I/O subsystem 1708, a storage subsystem 1718 and a communications subsystem 1724. Storage subsystem 1718 includes tangible computer-readable storage media 1722 and a system memory 1710.

Bus subsystem 1702 provides a mechanism for letting the various components and subsystems of computer system 1700 communicate with each other as intended. Although bus subsystem 1702 is shown schematically as a single bus, alternative embodiments of the bus subsystem may utilize multiple buses. Bus subsystem 1702 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. For example, such architectures may include an Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus, which can be implemented as a Mezzanine bus manufactured to the IEEE P1386.1 standard.

Processing unit 1704, which can be implemented as one or more integrated circuits (e.g., a conventional microprocessor or microcontroller), controls the operation of computer system 1700. One or more processors may be included in processing unit 1704. These processors may include single core or multicore processors. In certain embodiments, processing unit 1704 may be implemented as one or more independent processing units 1732 and/or 1734 with single or multicore processors included in each processing unit. In other embodiments, processing unit 1704 may also be implemented as a quad-core processing unit formed by integrating two dual-core processors into a single chip.

In various embodiments, processing unit 1704 can execute a variety of programs in response to program code and can maintain multiple concurrently executing programs or processes. At any given time, some or all of the program code to be executed can be resident in processor(s) 1704 and/or in storage subsystem 1718. Through suitable programming, processor(s) 1704 can provide various functionalities described above. Computer system 1700 may additionally include a processing acceleration unit 1706, which can include a digital signal processor (DSP), a special-purpose processor, and/or the like.

I/O subsystem 1708 may include user interface input devices and user interface output devices. User interface input devices may include a keyboard, pointing devices such as a mouse or trackball, a touchpad or touch screen incorporated into a display, a scroll wheel, a click wheel, a dial, a button, a switch, a keypad, audio input devices with voice command recognition systems, microphones, and other types of input devices. User interface input devices may include, for example, motion sensing and/or gesture recognition devices such as the Microsoft Kinect® motion sensor that enables users to control and interact with an input device, such as the Microsoft Xbox® 360 game controller, through a natural user interface using gestures and spoken commands. User interface input devices may also include eye gesture recognition devices such as the Google Glass® blink detector that detects eye activity (e.g., ‘blinking’ while taking pictures and/or making a menu selection) from users and transforms the eye gestures as input into an input device (e.g., Google Glass®). Additionally, user interface input devices may include voice recognition sensing devices that enable users to interact with voice recognition systems (e.g., Siri® navigator), through voice commands.

User interface input devices may also include, without limitation, three dimensional (3D) mice, joysticks or pointing sticks, gamepads and graphic tablets, and audio/visual devices such as speakers, digital cameras, digital camcorders, portable media players, webcams, image scanners, fingerprint scanners, barcode reader 3D scanners, 3D printers, laser rangefinders, and eye gaze tracking devices. Additionally, user interface input devices may include, for example, medical imaging input devices such as computed tomography, magnetic resonance imaging, position emission tomography, medical ultrasonography devices. User interface input devices may also include, for example, audio input devices such as MIDI keyboards, digital musical instruments and the like.

User interface output devices may include a display subsystem, indicator lights, or non-visual displays such as audio output devices, etc. The display subsystem may be a cathode ray tube (CRT), a flat-panel device, such as that using a liquid crystal display (LCD) or plasma display, a projection device, a touch screen, and the like. In general, use of the term “output device” is intended to include all possible types of devices and mechanisms for outputting information from computer system 1700 to a user or other computer. For example, user interface output devices may include, without limitation, a variety of display devices that visually convey text, graphics and audio/video information such as monitors, printers, speakers, headphones, automotive navigation systems, plotters, voice output devices, and modems.

Computer system 1700 may comprise a storage subsystem 1718 that provides a tangible non-transitory computer-readable storage medium for storing software and data constructs that provide the functionality of the embodiments described in this disclosure. The software can include programs, code modules, instructions, scripts, etc., that when executed by one or more cores or processors of processing unit 1704 provide the functionality described above. Storage subsystem 1718 may also provide a repository for storing data used in accordance with the present disclosure.

As depicted in the example in FIG. 17, storage subsystem 1718 can include various components including a system memory 1710, computer-readable storage media 1722, and a computer readable storage media reader 1720. System memory 1710 may store program instructions that are loadable and executable by processing unit 1704. System memory 1710 may also store data that is used during the execution of the instructions and/or data that is generated during the execution of the program instructions. Various different kinds of programs may be loaded into system memory 1710 including but not limited to client applications, Web browsers, mid-tier applications, relational database management systems (RDBMS), virtual machines, containers, etc.

System memory 1710 may also store an operating system 1716. Examples of operating system 1716 may include various versions of Microsoft Windows®, Apple Macintosh®, and/or Linux operating systems, a variety of commercially-available UNIX® or UNIX-like operating systems (including without limitation the variety of GNU/Linux operating systems, the Google Chrome® OS, and the like) and/or mobile operating systems such as iOS, Windows® Phone, Android® OS, BlackBerry® OS, and Palm® OS operating systems. In certain implementations where computer system 1700 executes one or more virtual machines, the virtual machines along with their guest operating systems (GOSs) may be loaded into system memory 1710 and executed by one or more processors or cores of processing unit 1704.

System memory 1710 can come in different configurations depending upon the type of computer system 1700. For example, system memory 1710 may be volatile memory (such as random access memory (RAM)) and/or non-volatile memory (such as read-only memory (ROM), flash memory, etc.) Different types of RAM configurations may be provided including a static random access memory (SRAM), a dynamic random access memory (DRAM), and others. In some implementations, system memory 1710 may include a basic input/output system (BIOS) containing basic routines that help to transfer information between elements within computer system 1700, such as during start-up.

Computer-readable storage media 1722 may represent remote, local, fixed, and/or removable storage devices plus storage media for temporarily and/or more permanently containing, storing, computer-readable information for use by computer system 1700 including instructions executable by processing unit 1704 of computer system 1700.

Computer-readable storage media 1722 can include any appropriate media known or used in the art, including storage media and communication media, such as but not limited to, volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage and/or transmission of information. This can include tangible computer-readable storage media such as RAM, ROM, electronically erasable programmable ROM (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disk (DVD), or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or other tangible computer readable media.

By way of example, computer-readable storage media 1722 may include a hard disk drive that reads from or writes to non-removable, nonvolatile magnetic media, a magnetic disk drive that reads from or writes to a removable, nonvolatile magnetic disk, and an optical disk drive that reads from or writes to a removable, nonvolatile optical disk such as a CD ROM, DVD, and Blu-Ray® disk, or other optical media. Computer-readable storage media 1722 may include, but is not limited to, Zip® drives, flash memory cards, universal serial bus (USB) flash drives, secure digital (SD) cards, DVD disks, digital video tape, and the like. Computer-readable storage media 1722 may also include, solid-state drives (SSD) based on non-volatile memory such as flash-memory based SSDs, enterprise flash drives, solid state ROM, and the like, SSDs based on volatile memory such as solid state RAM, dynamic RAM, static RAM, DRAM-based SSDs, magnetoresistive RAM (MRAM) SSDs, and hybrid SSDs that use a combination of DRAM and flash memory based SSDs. The disk drives and their associated computer-readable media may provide non-volatile storage of computer-readable instructions, data structures, program modules, and other data for computer system 1700.

Machine-readable instructions executable by one or more processors or cores of processing unit 1704 may be stored on a non-transitory computer-readable storage medium. A non-transitory computer-readable storage medium can include physically tangible memory or storage devices that include volatile memory storage devices and/or non-volatile storage devices. Examples of non-transitory computer-readable storage medium include magnetic storage media (e.g., disk or tapes), optical storage media (e.g., DVDs, CDs), various types of RAM, ROM, or flash memory, hard drives, floppy drives, detachable memory drives (e.g., USB drives), or other type of storage device.

Communications subsystem 1724 provides an interface to other computer systems and networks. Communications subsystem 1724 serves as an interface for receiving data from and transmitting data to other systems from computer system 1700. For example, communications subsystem 1724 may enable computer system 1700 to connect to one or more devices via the Internet. In some embodiments communications subsystem 1724 can include radio frequency (RF) transceiver components for accessing wireless voice and/or data networks (e.g., using cellular telephone technology, advanced data network technology, such as 3G, 4G or EDGE (enhanced data rates for global evolution), WiFi (IEEE 802.11 family standards, or other mobile communication technologies, or any combination thereof), global positioning system (GPS) receiver components, and/or other components. In some embodiments communications subsystem 1724 can provide wired network connectivity (e.g., Ethernet) in addition to or instead of a wireless interface.

In some embodiments, communications subsystem 1724 may also receive input communication in the form of structured and/or unstructured data feeds 1726, event streams 1728, event updates 1730, and the like on behalf of one or more users who may use computer system 1700.

By way of example, communications subsystem 1724 may be configured to receive data feeds 1726 in real-time from users of social networks and/or other communication services such as Twitter® feeds, Facebook® updates, web feeds such as Rich Site Summary (RSS) feeds, and/or real-time updates from one or more third party information sources.

Additionally, communications subsystem 1724 may also be configured to receive data in the form of continuous data streams, which may include event streams 1728 of real-time events and/or event updates 1730, that may be continuous or unbounded in nature with no explicit end. Examples of applications that generate continuous data may include, for example, sensor data applications, financial tickers, network performance measuring tools (e.g., network monitoring and traffic management applications), clickstream analysis tools, automobile traffic monitoring, and the like.

Communications subsystem 1724 may also be configured to output the structured and/or unstructured data feeds 1726, event streams 1728, event updates 1730, and the like to one or more databases that may be in communication with one or more streaming data source computers coupled to computer system 1700.

Computer system 1700 can be one of various types, including a handheld portable device (e.g., an iPhone® cellular phone, an iPad® computing tablet, a PDA), a wearable device (e.g., a Google Glass® head mounted display), a PC, a workstation, a mainframe, a kiosk, a server rack, or any other data processing system.

Due to the ever-changing nature of computers and networks, the description of computer system 1700 depicted in the figure is intended only as a specific example. Many other configurations having more or fewer components than the system depicted in the figure are possible. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, firmware, software (including applets), or a combination. Further, connection to other computing devices, such as network input/output devices, may be employed. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will appreciate other ways and/or methods to implement the various embodiments.

Although specific embodiments have been described, various modifications, alterations, alternative constructions, and equivalents are also encompassed within the scope of the disclosure. Embodiments are not restricted to operation within certain specific data processing environments, but are free to operate within a plurality of data processing environments. Additionally, although embodiments have been described using a particular series of transactions and steps, it should be apparent to those skilled in the art that the scope of the present disclosure is not limited to the described series of transactions and steps. Various features and aspects of the above-described embodiments may be used individually or jointly.

Further, while embodiments have been described using a particular combination of hardware and software, it should be recognized that other combinations of hardware and software are also within the scope of the present disclosure. Embodiments may be implemented only in hardware, or only in software, or using combinations thereof. The various processes described herein can be implemented on the same processor or different processors in any combination. Accordingly, where components are described as being configured to perform certain operations, such configuration can be accomplished, e.g., by designing electronic circuits to perform the operation, by programming programmable electronic circuits (such as microprocessors) to perform the operation, or any combination thereof. Processes can communicate using a variety of techniques including but not limited to conventional techniques for inter process communication, and different pairs of processes may use different techniques, or the same pair of processes may use different techniques at different times.

The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. It will, however, be evident that additions, subtractions, deletions, and other modifications and changes may be made thereunto without departing from the broader spirit and scope as set forth in the claims. Thus, although specific disclosure embodiments have been described, these are not intended to be limiting. Various modifications and equivalents are within the scope of the following claims.

The use of the terms “a” and “an” and “the” and similar referents in the context of describing the disclosed embodiments (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless otherwise indicated herein or clearly contradicted by context. The terms “comprising,” “having,” “including,” and “containing” are to be construed as open-ended terms (i.e., meaning “including, but not limited to,”) unless otherwise noted. The term “connected” is to be construed as partly or wholly contained within, attached to, or joined together, even if there is something intervening. Recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein and each separate value is incorporated into the specification as if it were individually recited herein. All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., “such as”) provided herein, is intended merely to better illuminate embodiments and does not pose a limitation on the scope of the disclosure unless otherwise claimed. No language in the specification should be construed as indicating any non-claimed element as essential to the practice of the disclosure.

Disjunctive language such as the phrase “at least one of X, Y, or Z,” unless specifically stated otherwise, is intended to be understood within the context as used in general to present that an item, term, etc., may be either X, Y, or Z, or any combination thereof (e.g., X, Y, and/or Z). Thus, such disjunctive language is not generally intended to, and should not, imply that certain embodiments require at least one of X, at least one of Y, or at least one of Z to each be present.

Preferred embodiments of this disclosure are described herein, including the best mode known for carrying out the disclosure. Variations of those preferred embodiments may become apparent to those of ordinary skill in the art upon reading the foregoing description. Those of ordinary skill should be able to employ such variations as appropriate and the disclosure may be practiced otherwise than as specifically described herein. Accordingly, this disclosure includes all modifications and equivalents of the subject matter recited in the claims appended hereto as permitted by applicable law. Moreover, any combination of the above-described elements in all possible variations thereof is encompassed by the disclosure unless otherwise indicated herein.

All references, including publications, patent applications, and patents, cited herein are hereby incorporated by reference to the same extent as if each reference were individually and specifically indicated to be incorporated by reference and were set forth in its entirety herein.

In the foregoing specification, aspects of the disclosure are described with reference to specific embodiments thereof, but those skilled in the art will recognize that the disclosure is not limited thereto. Various features and aspects of the above-described disclosure may be used individually or jointly. Further, embodiments can be utilized in any number of environments and applications beyond those described herein without departing from the broader spirit and scope of the specification. The specification and drawings are, accordingly, to be regarded as illustrative rather than restrictive.

Claims

1. A method, comprising:

accessing, by a virtual machine manager of a computing system, a source device configured to store device traffic in a source device guest memory;
configuring, by one or more worker nodes of a control plane of the computing system, a destination device;
transmitting, by the one or more worker nodes of the control plane of the computing device, a source device state to the destination device;
instructing, by the virtual machine manager of the computing device, the source device to pause storing the device traffic to the source device guest memory;
transmitting, by the virtual machine manager of the computing device, dirty page information to a destination device guest memory; and
instructing, by the virtual machine manager of the computing device, the destination device to initiate storing the device traffic to the destination device guest memory.

2. The method of claim 1, wherein the method further comprises:

limiting, by the virtual machine manager of the computing device, a bandwidth of device traffic to the source device.

3. The method of claim 2, wherein the bandwidth can be limited based at least in part on a queue depth of a source virtual machine (VM) queue or number of path groups.

4. The method of claim 1, wherein the destination device is configured by launching a virtual function that represents a nonvolatile memory express (NVMe) device on the destination device.

5. The method of claim 1, wherein a time period between instructing the source device to pause storing device traffic and instructing the destination device to resume storing device traffic is 500 milliseconds (ms) or less.

6. The method of claim 1, wherein transmitting the dirty page information further comprises:

storing, by the virtual machine manager of the computing device, dirty page information in a computing system memory; and
transmitting, by the virtual machine manager of the computing device, the dirty page information from the computing system memory to the destination device guest memory.

7. The method of claim 1, wherein transmitting the sources device state further comprises:

storing, by one or more worker nodes of a control plane of the computing system, a source device state in a computing system memory; and
transmitting, by the one or more worker nodes of the control plane of the computing device, the source device state from the computing system memory to the destination device.

8. A computer-program product tangibly embodied in one or more non-transitory machine-readable media, including instructions configured to cause one or more data processors to perform operations comprising:

accessing a source device configured to store device traffic in a source device guest memory;
configuring a destination device;
transmitting a source device state to the destination device;
instructing the source device to pause storing device traffic to the source device guest memory;
transmitting dirty page information to a destination device guest memory; and
instructing the destination device to initiate storing the device traffic to the destination device guest memory.

9. The computer-program product of claim 8, wherein the operations further comprise:

limiting a bandwidth of device traffic to the source device.

10. The computer-program product of claim 9, wherein the bandwidth can be limited based at least in part on a queue depth of a source virtual machine (VM) or a number of path groups.

11. The computer-program product of claim 8, wherein the destination device is configured by launching a virtual function that represents a nonvolatile memory express (NVMe) device.

12. The computer-program product of claim 8, wherein a time period between instructing the source device to pause storing device traffic and instructing the destination device to resume storing device traffic is 500 milliseconds (ms) or less.

13. The computer-program product of claim 8, wherein transmitting the dirty page information further comprises:

storing dirty page information in a computing system memory; and
transmitting the dirty page information from the computing system memory to the destination device guest memory.

14. The computer-program product of claim 8, wherein transmitting the sources device state further comprises:

storing a source device state in a computing system memory; and
transmitting the source device state from the computing system memory to the destination device.

15. A system comprising:

one or more data processors; and
one or more non-transitory computer readable media storing instructions which, when executed by the one or more data processors, cause the one or more data processors to perform operations comprising:
accessing a source device configured to store device traffic in a source device guest memory;
configuring a destination device;
transmitting a source device state to the destination device;
instructing the source device to pause storing device traffic to the source device guest memory;
transmitting dirty page information to a destination device guest memory; and
instructing the destination device to initiate storing the device traffic to the destination device guest memory.

16. The system of claim 15, wherein the operations further comprise:

limiting a bandwidth of device traffic to the source device.

17. The system of claim 16, wherein the bandwidth can be limited based at least in part on a queue depth of a source virtual machine (VM) queue or a number of path groups.

18. The system of claim 15, wherein the destination device is configured by launching a virtual function device that represents a nonvolatile memory express (NVMe) device.

19. The system of claim 15, wherein a time period between instructing the source device to pause storing device traffic and instructing the destination device to resume storing device traffic is 500 milliseconds (ms) or less.

20. The system of claim 15, wherein the dirty page information includes device traffic that is sent to the source device guest memory while storing the device traffic is paused.

Patent History
Publication number: 20240143370
Type: Application
Filed: Oct 28, 2022
Publication Date: May 2, 2024
Applicant: Oracle International Corporation (Redwood Shores, CA)
Inventors: Kent Lee (Ladera Ranch, CA), Jinlong Cai (Redmond, WA), Qiang Wang (Bellevue, WA)
Application Number: 17/976,476
Classifications
International Classification: G06F 9/455 (20060101);