HCIP-CloudServiceSolutionsArchitectV3 0TrainingMaterial
HCIP-CloudServiceSolutionsArchitectV3 0TrainingMaterial
and terminals.
• With the advent of personal computers (PCs) in the 1980s, the second platform
emerged, which was characterized by the client/server system, Ethernet, RDBMS,
and Web applications.
• Today we are using the third platform, which includes cloud computing, big data,
mobile devices, and socialization technologies. At the core of these technologies
is cloud computing. Customers use cloud providers' services to allocate IT
resources. Big data turbocharges data analysis to achieve in-depth insights and
for leaders' to make better-informed decisions. Mobile devices enable ubiquitous
access to applications and information. Socialization technologies help connect
people and ensure better collaboration and information exchanges.
• Agility and resource scheduling are embodied in the previous page. This page
focuses on cloud service enablement. That is, the new capabilities that are
offered in services.
• The cloud computing software industry has been a national priority since the
12th Five-Year Plan.
• According to the 13th Five-Year Science and Technology Innovation Plan, cloud
computing technologies and applications will be promoted to empower the next
generation of ICT infrastructure.
• In the Outline of the 14th Five-Year Plan for National Economic and Social
Development and Long-Range Objectives Through the Year 2035, the
development of StatChina was propelled to new heights and cloud computing
has becoming key to that growth. Cloud computing software will embrace new
opportunities.
• Building a cloud-based software system is very similar to building a house. If the
foundation is not solid, structural problems may damage the integrity and
functionality of the house. When designing a solution for migrating enterprise
applications to the cloud, if you ignore security, reliability, scalability,
performance, and cost optimization, it may be difficult to build a system that
meets your expectations and requirements. Considering the following factors in
the design will help you build a stable and efficient system:
• Scalability: The system can be scaled out or scaled up according to the number of
users or overall workload.
• Tenants are also responsible for the security management of any application
software or utility they deploy on Huawei Cloud. Before deploying security
workloads in the production environment, tenants should test these workloads to
prevent adverse effects on their applications and services.
• Tenants own and control their data regardless of the Huawei Cloud service they
use. Tenants take measures to guarantee data confidentiality, integrity, and
availability, as well as the identity authentication and authorization for data
access. For example, tenants using IAM and DEW need to configure rules to
properly keep their own service login accounts, passwords, and keys.
• The longest annual downtime allowed for each SLA level is calculated as follows
(365 days in a year):
• An annual downtime of 5.26 minutes means 99.999% SLA. A better SLA means
higher requirements on the system. As a result, we need to consider whether the
system is capable of meeting the increasing SLA requirements.
• The common cloud system HA design solutions are as follows:
▫ The on-premises HA solution applies to on-premises production centers and
single-AZ scenarios.
▫ The intra-city HA/DR solutions, including an active-active data center
solution and an active-passive DR solution, apply to the HA design of intra-
city DR centers and dual-AZ scenarios.
▫ The remote HA/DR solutions, including a geo-redundant DR solution and
an active-passive DR solution, apply to remote DR centers and cross-region
HA.
• Prevention of performance bottlenecks
• A transaction bill includes the billing information of each order and of each
billing cycle (a cloud service billing cycle can be hourly, daily, or monthly).
• High service flexibility – scalability
• Static content refers to the same content obtained through different access
requests, such as images, videos, and file packages on websites. CDN can provide
acceleration services for static content under acceleration domain names.
• CAPEX indicates the capital expenditure, such as fund and fixed assets. For
example, the once-off expenditure on network equipment, computers, and
instruments is CAPEX, among which network equipment accounts for the largest
proportion.
• Answer: False. After migrating workloads to the cloud, organizations need to
adopt cost design. If cloud resources are used without restrictions, the cost will
far exceed that of the off-cloud architecture.
▫ Pay-as-you-go: You only need to pay for what you use by the hour or even
by the minute.
▫ ECSs are deployed in different AZs, so that if one AZ becomes faulty, other
AZs in the same region will not be affected.
▫ Cloud Eye lets you keep a close eye on the performance and resource
utilization of ECSs, ensuring their reliability and availability.
• General computing-basic
▫ Suitable for scenarios that require moderate CPU performance generally
but occasionally burstable high performance while keeping costs low
• General computing:
▫ Suitable for websites and web applications, small-scale databases and
cache servers, and light- and medium-workload enterprise applications with
strict requirements on PPS
• General computing-plus
▫ Suitable for heavy- and medium-load enterprise applications that have
higher requirements on computing and network performance, such as web
applications, e-commerce platforms, short video platforms, online games,
insurance, and finance
• Memory-optimized
▫ Suitable for massive parallel processing (MPP) data warehouses,
MapReduce and Hadoop distributed computing, distributed file systems,
network file systems, and log or data processing applications
• Disk-intensive
▫ Suitable for distributed file systems, network file systems, and log or data
processing applications
• High-performance computing
▫ Computing and storage systems for genetic engineering, games,
animations, and biopharmaceuticals
• The types displayed in the table were current as of the end of August 2022.
• To select the right ECS type, consider the following factors:
▫ Service deployment: Deploy ECSs in the region closest to your services to
reduce network delay and improve the access speed.
▫ Resource utilization: Make full use of purchased cloud resources. Do not
buy more capacity than is needed.
▫ Specification adjustment: In the subsequent content, we'll examine a
hypothetical startup to look at how to select the right ECS types for
different development stages (startup, growth, and maturity).
▫ Cost control: Selecting the right ECS types and specifications help control
costs. Evaluate your service scale and budget and scale up ECSs or change
ECS types to meet service demands.
• T6 family:
▫ The performance of general-computing basic T6 ECSs is restricted by the
benchmark performance and CPU credits.
▫ Suitable for scenarios where the CPU usage is low but requires burstable
CPU power, for example, microservices.
• S6 family:
▫ S6 ECSs are equipped with second-generation Intel® Xeon® Scalable
processors and Huawei 25GE high-speed intelligent NICs that cost-
effectively provide high network bandwidth and PPS throughput.
▫ Suitable for websites and web applications with high requirements for PPS
• S7 family:
• What is PPS?
▫ PPS, short for packets per second, is the number of network data packets
that can be processed by an ECS per second, including the number of sent
and received packets, including both private and public traffic. The
maximum PPS is the maximum number of data packets an ECS can process,
both incoming and outgoing per second.
• C3 family:
▫ C3 ECSs use Intel® Xeon® Scalable processors to provide high and stable
computing performance. Working in high-performance networks, the C3
ECSs deliver higher performance and stability, meeting enterprise-class
application requirements.
▫ Suitable for small- and medium-sized databases, cache clusters, and search
clusters that have high requirements on stability.
• C6s family:
▫ Suitable for Internet, gaming, and rendering scenarios, especially those that
require high computing and network stability.
• C7 family:
▫ C7 ECSs use third-generation Intel® Xeon® Scalable processors to provide
enhanced compute, security, and stability. A C7 ECS can be configured with
up to 128 vCPUs and 3,200 MHz memory. C7 ECSs support secure reboot
and provide secure, trusted cloud environment for applications to run in.
▫ Suitable for heavy- and medium-load enterprise applications that demand
more compute and network performance, such as web applications, e-
commerce platforms, short video platforms, online games, insurance, and
finance applications.
• M7 family:
• D7 family:
▫ D7 ECSs are mainly used for massively parallel processing (MPP) data
warehouses, MapReduce and Hadoop distributed computing, and big data
computing.
▫ Suitable for distributed file systems, network file systems, and log or data
processing applications.
• I7 family:
▫ I7 ECSs use high-performance local NVMe SSDs to provide high IOPS and
low read/write latency.
▫ Suitable for high-performance relational databases, non-relational
databases, and ElasticSearch search.
• ECSs should be continuously optimized.
• Billing modes: yearly/monthly, pay-per-use, and spot price
▫ Yearly/monthly: You can purchase a yearly/monthly ECS subscription and
enter your required duration. Yearly/monthly subscriptions are pre-paid
with a single, lump sum payment.
▫ Pay-per-use: You do not need to set a required duration after setting ECS
configurations. The system bills your account based on the service duration.
▫ Spot price: Huawei Cloud sells available spare compute resources at a
discount. The price changes in real time depending on market supply and
demand.
• Region and AZ: ECSs in different regions cannot communicate with each other
over a private network. Select a region closest to your services to ensure low
network latency and quick access.
• Specifications: A broad set of ECS types are available for you to choose from. You
can choose from existing types and flavors in the list, or enter a flavor or specify
vCPUs and memory size to search for the flavor suited to your needs.
• Image: An image is a server or disk template that contains an OS or service data
and necessary application software. IMS provides public, private, Marketplace,
and shared images.
• System disk types: high I/O, general-purpose SSD, ultra-high I/O, and extreme
SSD. By default, you need to specify the type and size of the system disk.
• Network settings for an ECS:
▫ Subnet: A subnet is a range of IP addresses in your VPC and provides IP
address management and DNS resolution functions for ECSs in it. The IP
addresses of all ECSs in a subnet belong to the subnet.
▫ Security group: A security group is a collection of access control rules for
ECSs that have the same security protection requirements and that are
mutually trusted. It helps to enhance ECS security.
▫ Extension NIC: optional
• Advanced settings for an ECS:
▫ ECS name: You can customize ECS names in compliance with naming rules.
If you intend to purchase multiple ECSs at a time, the system automatically
adds a hyphen followed by a four-digit incremental number to the end of
each ECS.
▫ Login mode: Key pair allows you to use a key pair for login authentication.
Password allows you to use a username and its initial password for login
authentication. For Linux ECSs, the initial password is the root password.
For Windows ECSs, the initial password is the Administrator password.
▫ Cloud Backup and Recovery: With CBR, you can back up data for ECSs and
EVS disks, and use backups to restore the ECSs and EVS disks when
necessary.
▫ ECS group (Optional): An ECS group allows ECSs within the group to be
automatically allocated to different hosts.
▫ Advanced options: You can configure other advanced and optional settings.
• Cost-effectiveness: DeH allows you to bring your own license (BYOL), such as
licenses for Microsoft Windows Server, Microsoft SQL Server, and Microsoft
Office.
• Flexibility: You can apply for your DeHs flexibly. Your DeHs will be allocated
within several minutes.
• Application scenario: If you do not use the ECSs deployed on a DeH or want to
delete them after a period of time, you can migrate the ECSs to a public resource
pool.
• High security and reliability:
▫ BMS provides you with dedicated computing resources. You can add servers
to VPCs and security groups for network isolation and integrate related
components for server security. BMSs run on a QingTian architecture and
can use EVS disks, which can be backed up for restoration. BMS
interconnects with Dedicated Storage Service (DSS) to ensure the data
security and reliability required by enterprise services.
• High performance:
▫ BMS has no virtualization overhead, so the compute resources are fully
dedicated to running services. The QingTian they run on, an architecture
from Huawei, is designed with hardware-software synergy in mind. BMS
supports high-bandwidth, low-latency storage and networks on the cloud,
meeting the deployment density and performance requirements of mission-
critical services such as enterprise databases, big data, containers, HPC, and
AI.
• Agile deployment:
▫ The hardware-based acceleration provided by the QingTian architecture
enables EVS disks to be used as system disks. The required BMSs can be
provisioned within minutes of when you submit your order. You can
manage your BMSs throughout their lifecycle from the management
console or using open APIs with SDKs.
• Quick integration:
▫ BMSs can easily cooperate with the other cloud resources in a VPC, just like
ECSs do, to run a variety of cloud solutions (such as databases, big data
applications, containers, HPC, and AI solutions), accelerating cloud
transformation.
• VPC:
• User-defined VLAN:
▫ User-defined VLAN NICs are deployed in pairs. You can configure NIC bonds
to ensure high availability. User-defined VLANs in different AZs cannot
communicate with each other.
• Database:
• Big data:
▫ Internet services involving big data storage and analysis. BMS uses a
decoupled storage and compute solution that combines local storage and
Object Storage Service (OBS).
• Container:
• HPC/AI:
• Cloud phones provide video, audio, and touch SDKs. You can develop applications
based on terminals to obtain audios and videos of cloud phones. Alternatively,
you can collect touch instructions, for example, touch, slide, or click instructions,
and execute them on cloud phones.
• Easy-to-use:
▫ Deployment and O&M of containerized applications can be automated and
performed all in one place throughout the application lifecycle.
▫ Helm charts are pre-integrated, delivering out-of-the-box usability.
• High performance:
▫ CCE draws on years of field experience in compute, networking, storage,
and heterogeneous infrastructure. You can concurrently launch containers
at scale.
▫ The bare-metal NUMA architecture and high-speed InfiniBand network
cards yield three- to five-fold improvement in computing performance.
• Secure and reliable:
▫ CCE allows you to deploy nodes and workloads in a cluster across AZs. Such
a multi-active architecture ensures service continuity against host faults,
data center outages, and natural disasters.
▫ Clusters are private and completely controlled by users with deeply
integrated IAM and Kubernetes RBAC. You can set different RBAC
permissions for IAM users on the console.
• Open and compatible:
▫ CCE streamlines deployment, resource scheduling, service discovery, and
dynamic scaling of applications that run in Docker containers.
▫ CCE is built on Kubernetes and compatible with Kubernetes native APIs and
kubectl (a command line tool). CCE provides full support for the most
recent Kubernetes and Docker releases.
• When using FunctionGraph, you do not need to apply for or pre-configure any
compute, storage, or network services. Simply upload and run code in supported
runtimes. FunctionGraph provides and manages underlying compute resources,
including CPUs, memory, and networks. It also supports configuration and
resource maintenance, code deployment, automatic scaling, load balancing,
secure upgrade, and resource monitoring.
• Convenient: You can use a public, Marketplace, or private image to create ECSs in
batches, simplifying service deployment. You can also share, replicate, or export
images between different accounts, regions, or even cloud platforms.
• Secure: To ensure data reliability and durability, multiple copies of image files are
stored using Object Storage Service (OBS). You can use the envelope encryption
provided by Key Management Service (KMS) to encrypt private images.
• Flexible: You can manage the lifecycle of images using the management console
or APIs as needed. IMS can meet your requirements no matter you want to
migrate servers to the cloud, back up server environments, or migrate servers
between different accounts or regions on the cloud.
▫ A data disk image contains only service data. You can use a data disk image
to create EVS disks and use them to migrate your service data to the cloud.
• Constraints:
▫ You cannot share private images that have been published in Marketplace.
▫ You can share images only within a given region. To share an image across
regions, you need to replicate the image to the target region first.
▫ A system disk image or data disk image can be shared with a maximum of
128 users, and a full-ECS image can be shared with a maximum of 10 users.
▫ Only full-ECS images created from an ECS or a CBR backup can be shared.
• When you submit a request for creating a full-ECS image from an ECS, the
system will automatically create a backup for the ECS and then use the backup
to create a full-ECS image.
• The time required for creating a full-ECS image depends on the disk size,
network quality, and the number of concurrent tasks.
• The ECS used to create a full-ECS image must be in Running or Stopped state. To
create a full-ECS image containing a database, use a stopped ECS.
• When a full-ECS image is being created, if you detach the system disk from the
ECS or stop, start, or restart the ECS, the image creation will fail.
• If there are snapshots of the system disk and data disks but the ECS backup
creation is not complete, the full-ECS image you create will only be available in
the AZ where the source ECS is and can only be used to provision ECSs in this AZ.
You cannot provision ECSs in other AZs in the region until the original ECS is fully
backed up and the full-ECS image is in the Normal state.
• If you use a full-ECS image to change an ECS OS, only the system disk data can
be written into the ECS. Therefore, if you want to restore or migrate the data disk
data of an ECS by using a full-ECS image, you can only use the image to create a
new ECS rather than use it to change the ECS OS.
• When there are more resources available than what is needed, idle resources are
wasted.
• When there are not enough resources available, user experience deteriorates.
User churn increases and revenue is lost.
• Automatic scaling helps you automatically meet customer requirements.
• The process of using AS is as follows:
• AS advantages:
▫ Improved availability: With AS, your applications always have the right
amount of resources at the right time. When working with ELB, AS
automatically associates a load balancing listener with any instances newly
added to an AS group. Then, ELB automatically distributes access traffic to
all healthy instances in the AS group through the listener.
▫ In the example shown here, when the number of access requests reaches
1,000, the existing resources cannot handle the demand. More resources
are needed. When the peak hours pass, idle resources need to be removed
to avoid waste and reduce costs.
▫ AS can work together with Cloud Eye to do this automatically. When Cloud
Eye detects resources reach a threshold you have specified in an AS policy,
for example, CPU usage higher than 70%, memory usage higher than 80%,
or access requests more than 500, AS automatically triggers scaling actions
to add more resources.
• When you use AS, you need to create an AS group, create an AS configuration,
and then configure an AS policy for the AS group.
• Then AS checks whether the condition specified in the AS policy is met, and
determines whether to execute a scaling action based on the results.
• An AS policy can trigger scaling actions to scale ECS and bandwidth resources for
an AS group. An AS policy defines the conditions for triggering a scaling action
and the operation that will be performed. When the condition is met, a scaling
action is triggered automatically. AS supports alarm-based, scheduled, and
periodic scaling policies.
▫ You manually increase the expected number of instances for the AS group
or AS automatically adds instances to the AS group.
• The instance status changes from Enabled to Removing from AS group when any
of the following occurs:
▫ You manually decrease the expected number of instances for the AS group
or the system automatically removes instances from the AS group.
• 2. Answer: ABC
• Discussion 1:
• Discussion 2:
▫ Security: dynamic and static data security, network security, and access
control
• Users and applications with high mobility require the flexibility and scale of cloud
networks for assured performance, security, and easier management.
• Cloud networks also improve IT efficiency and save money for offices, schools,
home office, healthcare, and public spaces.
• There are the following types of network services:
• Cloud networks:
▫ General networks and security policies: VPCs, security groups, and network
ACLs
▫ Communications within a given region on the cloud: VPC Endpoint and VPC
Peering
▫ Ensure that the VPC CIDR block does not overlap with the enterprise private
network. If there are multiple VPCs in different regions, the VPC CIDR
blocks cannot overlap.
▫ Do not allocate all subnets and IP addresses at once. You should reserve
space for future capacity expansion.
▫ Select private CIDR blocks for VPCs and subnets, which are used for private
communications. If a public CIDR block is configured, conflicts may occur
during internet access.
• Security group 2: The first rule allows the App server to communicate with other
App servers that may be added later for capacity expansion. The second rule
allows Web 1 server to access the App server. The third rule allows all outbound
traffic.
• Network ACL 1: The first rule denies the access from the test subnet. The second
rule allows all inbound access, excepting the access from the test subnet denied
by the first rule. The third rule allows all outbound traffic.
• Network ACL 2: The first rule denies the access from the production subnet. The
second rule allows all inbound access, excepting the access from the production
subnet denied by the first rule. The third rule allows all outbound traffic.
• If two VPCs connected by a VPC peering connection overlap with each other,
there will be route conflicts and the VPC peering connection may not be usable.
• If there are three VPCs, A, B, and C, and VPC A is peered with both VPC B and
VPC C, but VPC B and VPC C overlap with each other, you cannot configure
routes with the same destinations for VPC A.
• You cannot have more than one VPC peering connection between the same two
VPCs at the same time.
• VPC peering does not support transitive peering relationships. For example, if VPC
A is connected to both VPC B and VPC C, but VPC B and VPC C are not
connected, VPC B and VPC C cannot communicate with each other through VPC
A. You need to create a VPC peering connection between VPC B and VPC C.
• A VPC peering connection between VPCs in different regions will not be usable.
• If you request a VPC peering connection with a VPC of another account, the
connection takes effect only after the peer account accept the request. If you
request a VPC peering connection with a VPC of your own, the system
automatically accepts the request and activates the connection.
VPCEP provides two types of resources: VPC endpoint services and VPC endpoints.
• VPC endpoint services refer to cloud services or your private services that can be
configured in VPCEP to provide services to users. For example, you can create an
application in a VPC and configure it as a VPC endpoint service that VPCEP
supports.
• VPC endpoints are channels for connecting VPCs to VPC endpoint services. You
can create an application in your VPC and configure it as a VPC endpoint service.
A VPC endpoint can be created in another VPC in the same region and then used
as a channel to access the VPC endpoint service.
• Function:
• Access scenario:
▫ VPC peering connections, in most cases, are used to connect subnets of two
VPCs belong to the same tenant.
• With VPC endpoint 1, the user's on-premises data center can access ELB in VPC 1.
• With VPC endpoint 2, the user's on-premises data center can access Elastic Cloud
Servers (ECSs) in VPC 2.
• With VPC endpoint 3, the user's on-premises data center can access Domain
Name Service (DNS) over the intranet.
• With VPC endpoint 4, the user's on-premises data center can access Object
Storage Service (OBS) over the intranet.
• High data security
▫ Huawei hardware uses IKE and IPsec to encrypt data to provide carrier-class
reliability and ensure a stable VPN connection.
• Seamless scale-out
▫ With VPN, you can connect your local data center to your VPC and quickly
extend services at the data center to the cloud, thereby forming a hybrid
cloud architecture.
• The connection is a dedicated network connection between your premises and a
Direct Connect location over a line you lease from a carrier. You can create a
standard connection by yourself or request a hosted connection from a partner.
After you are certified as a partner, you can also create an operations connection.
▫ A hosted connection allows you to share a port with others. Partners with
operations connections can provision hosted connections and allocate
VLANs and bandwidths for those connections. Only one virtual interface
can be created for each hosted connection.
• The virtual gateway is a logical gateway for accessing VPCs. Each VPC can have
only one virtual gateway associated, but multiple connections can use the same
virtual gateway to access one VPC.
• The virtual interface links a connection with one or more virtual gateways, each
of which is associated with a VPC, so that your on-premises network can
communicate with all these VPCs.
• VPN
• Direct Connect
▫ Low and stable latency, low jitter level, and excellent performance
Prerequisites:
▫ Auto-negotiation for the port has been disabled. The port speed and full-
duplex mode have been manually configured.
▫ Your device supports Border Gateway Protocol (BGP) and does not use
Autonomous System Number (ASN) 64512, which is used by Huawei Cloud.
Constraints:
▫ If you load a VPC to a cloud connection created using the same account,
you cannot enter loopback addresses, multicast addresses, or broadcast
addresses for the custom CIDR block.
▫ If a NAT gateway has been created for any VPC you have loaded to a cloud
connection, a custom CIDR block needs to be added and set to 0.0.0.0/0.
• Shared bandwidth:
▫ Shared bandwidth allows ECSs, BMSs, and load balancers that are bound
with EIPs from the same region to share the same bandwidth.
▫ When you host a large number of applications on the cloud, if each EIP
uses an independent bandwidth, a lot of bandwidths are required, which
significantly increases bandwidth costs. If all EIPs share the same
bandwidth, you can lower bandwidth cost and easily perform O&M.
• Dynamic BGP:
▫ Dynamic BGP provides automatic failover and chooses the best path based
on real-time network conditions and preset policies.
• Static BGP:
• Comparison in assurance:
• Dynamic BGP:
▫ When a fault occurs on a carrier's link, dynamic BGP will quickly select
another path to take over services, ensuring service availability.
▫ Currently, carriers in China that support dynamic BGP routing include China
Telecom, China Mobile, China Unicom, China Education and Research
Network (CERNET), National Radio and Television Administration, and Dr.
Peng Group.
• Static BGP:
▫ When changes occur on a network that uses static BGP, the manual
configuration takes some time and high availability cannot be guaranteed.
• Dedicated load balancers give you exclusive access to their resources, so the
performance of a dedicated load balancer is not affected by other load balancers.
In addition, there are a wide range of specifications available for selection.
• Shared load balancers are deployed in clusters, where all the load balancers
share resources. With a shared load balancer, the performance of one load
balancer can be affected by other load balancers.
• ELB periodically sends heartbeat messages to associated backend servers to
check their health to ensure that traffic is distributed only to healthy backend
servers. This can improve the availability of applications.
• The maximum stickiness duration at Layer 7 is 24 hours.
▫ For an application that has predictable peaks and troughs in traffic volumes,
ELB works with Auto Scaling to add or remove backend servers to keep up
with the changing demand. ELB routes requests to the required number of
backend servers to handle the load of your application based on the load
balancing algorithm and health check you set. One example is flash sales,
during which application traffic spikes in a short period. ELB can work with
Auto Scaling to run only the required number of backend servers, helping
to minimize IT costs.
• Cross-AZ load balancing:
▫ For services that require high availability, ELB can distribute traffic across
AZs. If an AZ becomes faulty, ELB distributes the traffic to backend servers
in other AZs that are running properly.
▫ ELB is ideal for banking, policing, and large application systems that require
high availability.
• Flexible deployment
▫ A public NAT gateway can be shared across subnets and AZs, so that even
if an AZ fails, the public NAT gateway can still run normally in another AZ.
The type and EIP of a public NAT gateway can be changed at any time.
• Ease of use
▫ Multiple types of public NAT gateways are available. Public NAT gateway
configuration is simple, the O&M is easy, and they can be provisioned
quickly. Once provisioned, they are stable and reliable.
• Cost-effectiveness
▫ With a public NAT gateway, when you send data through a private IP
address or provide services accessible from the Internet, the public NAT
gateway translates the private IP address to a public IP address. You no
longer need to configure one EIP for each server, which saves money on
EIPs and bandwidth.
• Transit subnet: A transit subnet is where a transit IP address resides.
• Strong security
• Easy O&M
▫ A private NAT gateway can map the CIDR block of each department to the
same VPC CIDR block, which simplifies the management of complex
networks.
• Zero conflicts
▫ You can migrate an in-use website domain name to the DNS service. To
ensure that your website services are not interrupted during the migration,
we will create a public zone and add DNS record sets for your website in
advance.
• Public domain name resolution: maps domain names to public IP addresses so
that your users can access your website or web applications over the Internet. A
public zone contains information about how a domain name and its subdomains
are translated into IP addresses for routing traffic over the Internet.
• Private domain name resolution: Translates private domain names into private IP
addresses to facilitate access to cloud resources within VPCs. A private zone
contains information about how to map a domain name (such as ecs.com) and
its subdomains used within one or more VPCs to private IP addresses (such as
192.168.1.1). With private domain names, your ECSs can communicate with each
other within the VPCs without having to connect to the Internet. These ECSs can
also access cloud services, such as OBS and Simple Message Notification (SMN),
over a private network.
• Intelligent resolution: returns different resolution results for the same domain
name based on the carrier networks or geographic locations of user IP addresses.
For example, if the visitor is a China Unicom user, the DNS server will return an
IP address of China Unicom. With this function, you can improve DNS resolution
efficiency and speed up cross-network access. You can also create more fine-
grained resolution lines based on source IP addresses.
• ABC
• A
• Discussion 1:
• Discussion 2:
▫ To reduce costs, delete servers that are not working in a backend server
group for load balancing immediately.
• DAS: Although DAS is old, it is still suitable for scenarios where the data volume
is small and the requirement for access speed is not high.
• NAS: NAS is suitable for file servers to store unstructured data. Although their
access speed is limited by the Ethernet, NAS can be flexibly deployed at low costs.
• SAN: SAN is suitable for large-scale applications or database systems. But SAN is
costly and complex.
• Block storage: Block storage breaks up data into blocks and then stores those
blocks as separate pieces, each with a unique identifier. Those blocks of data can
be placed wherever it is most efficient. That means each block can be configured
(or partitioned) to work with different operating systems.
• File storage: File storage is also referred to as file-level or file-based storage. File
storage data is stored as single pieces of data in folders.
• Recovery Time Objective (RTO): the maximum tolerable service downtime, from
the time when a disaster happened to the time when services were recovered
• Precautions:
▫ When attaching a disk, ensure that the server and disk reside in the same
AZ. Or, the attachment will fail.
▫ You can create snapshots for disks regularly and use snapshots to recover
your data in case that data is lost or inconsistent due to misoperations,
viruses, or attacks.
▫ You can use a snapshot to create multiple disks containing the same initial
data, and these disks can be used as data resources for various services.
• A bucket is a container for storing objects in OBS. OBS offers a flat structure
based on buckets and objects. This structure enables all objects to be stored at
the same logical layer, rather than being stored hierarchically. Each bucket has its
own properties, such as the storage class, access control, and region. You can
create buckets with required storage classes and access control in different
regions and further configure advanced settings, to meet storage requirements in
a wide range of scenarios.
• OBS provides massive storage for files of any format, catering to the needs of
common users, websites, enterprises, and developers. Neither the entire OBS
system nor any single bucket has limitations on the storage capacity or the
number of objects/files that can be stored. As a web service, OBS supports APIs
over HTTP and HTTPS. You can easily access and manage data stored in OBS
anytime, anywhere through OBS Console or OBS tools. In addition, OBS SDKs and
APIs make it easy to manage data stored in OBS and to develop upper-layer
applications.
• Standard:
• Infrequent Access:
▫ The Infrequent Access storage class can be used for file synchronization and
sharing, enterprise backups, and many other scenarios. It has the same
durability, low latency, and high throughput as the Standard storage class,
with a lower cost, but its availability is slightly lower than the Standard
storage class.
• Archive:
▫ The Archive storage class is ideal for scenarios such as data archive and
long-term backups. It is secure and durable and delivers the lowest cost
among the three storage classes. The OBS Archive storage class can be
used to replace tape libraries. To save money, it may take hours to restore
the archived data.
• You can choose multi-AZ storage or single-AZ storage as your redundancy policy
based on your business needs. The multi-AZ storage stores data in multiple AZs
to deliver up to 99.9999999999% of data durability and up to 99.995% of service
continuity, far higher than those of a conventional architecture.
• The 12 nines of durability means that the average annual loss rate of objects is
expected to be 0.0000000001%. For example, if you store 100 million objects in
OBS, only one object may be lost every 10,000 years.
▫ There is no limit on the number of lifecycle rules in a bucket, but the total
size of XML descriptions about all lifecycle rules in a bucket cannot exceed
20 KB.
• There are some restrictions on storage class transition using lifecycle rules:
▫ Lifecycle rules can transition objects only from the Standard storage class to
Infrequent Access storage class, or from the Standard or Infrequent Access
storage class to Archive storage class.
▫ If you want to change the storage class back from Infrequent Access to
Standard, or from Archive to Standard or Infrequent Access, you must
manually transition the storage class. In addition, to change the storage
class of an archived object, you need to manually restore the object first.
• You can configure a rule to replicate only objects with a specified prefix or to
replicate all objects in a bucket. Replicated objects in the destination bucket are
copies of those in the source bucket. Objects in both buckets have the same
names, metadata, content, sizes, last modification time, creators, version IDs,
user-defined metadata, and ACLs. By default, a source object and its copy have
the same storage class, but you can also specify a different storage class for an
object copy if you want.
▫ Updated objects, for example, the object content is updated or the copied
ACL is updated.
▫ With this method, you need to create a key using Key Management Service
(KMS) or use the default key provided by KMS. The KMS key is then used
for server-side encryption when you upload objects to OBS.
▫ For this method, the customer-provided keys and their MD5 values are used
for server-side encryption.
• Events supported by OBS are listed as follows:
• OBS provides APIs such as PUT, POST, and COPY for uploading objects. You can
configure event types corresponding to these APIs. Then, when you use such an
API to upload an object, you will receive a notification. You can also configure the
ObjectCreated:* event type to obtain all object upload notifications.
▫ ObjectCreated:* (all upload operations)
▫ ObjectCreated:Put (uploading an object)
▫ ObjectCreated:Post (uploading an object with a browser)
▫ ObjectCreated:Copy (copying an object)
▫ ObjectCreated:CompleteMultipartUpload (merging parts)
• By configuring the ObjectRemoved event type, you can receive a notification
when one or more objects are removed from a bucket.
• By configuring the ObjectRemoved:Delete event type, you can receive a
notification when an object is deleted or an object version is permanently deleted.
By configuring the ObjectRemoved:DeleteMarkerCreated event type, you can
receive a notification when a delete marker is added to an object. You can also
use ObjectRemoved:* to receive a notification each time an object is deleted.
▫ ObjectRemoved:* (all delete operations)
▫ ObjectRemoved:Delete (deleting an object)
▫ ObjectRemoved:DeleteMarkerCreated (adding a delete marker to an object)
• The OBS big data solution is designed for a variety of scenarios, including storage
and analysis of massive amounts of data, query of historical data details, analysis
of a large number of behavior logs, and analysis and statistics of public
transactions.
▫ Storage for petabytes of data, batch data analysis, and response for data
detail queries in seconds
▫ Analysis of learning habits and operation logs, as well as analysis and query
of system operation logs
• SFS: Like a shared folder, for example, a remote shared directory in Windows, the
file system already exists, and you can directly store data to the file system.
• OBS: Each piece of data corresponds to a unique ID. Object storage does not
have the directory structure similar to file storage. Data is stored in a flat
structure, and you can locate data by object ID.
• Various specifications:
▫ High I/O storage is suitable for scenarios that require high performance,
high read/write speed, and real-time data storage.
• Elastic scalability:
▫ Linear performance scaling: DSS disks can be expanded while services are
running, and linear performance increase can be achieved.
▫ Both system disks and data disks can be encrypted for improved data
security.
▫ CBR allows you to create backups for DSS disks and restore the disk data
using backups. Backups can be created for a DSS disk, maximizing data
security and integrity and ensuring service security.
• Enterprise customers: IDC hosting customers, securities settlement companies,
and more.
• Customers use EVS shared storage and DSS dedicated storage for their services.
EVS provides storage for enterprise OA, development and testing, and databases.
DSS provides storage for the mission-critical services running on BMSs.
• CDN facilitates whole network access across carriers and regions. Websites
cannot be accessed due to various factors, such as regional ISP limitation and
egress bandwidth limitation. CDN can cover global lines. It cooperates with
carriers to deploy Internet Data Center resources and edge nodes on networks of
backbone node providers. CDN helps customers make the most of bandwidth
resources and balance origin server traffic.
• Load balancing and distributed storage of CDN enhance website security and
reliability to cope with most Internet attacks. The anti-attack system can also
protect websites from malicious attacks.
• CDN supports remote backups. When a server is faulty, the system switches
services to other adjacent healthy server nodes. The reliability is close to 100%,
and websites never breaks down.
• With CDN, customers can delivery content to global users without worrying
about server investments, subsequent hosting and O&M, image synchronization
between servers, or O&M personnel. CDN helps customers save human, energy,
and financial resources.
• CDN enables customers to stay focused on their core services. CDN vendors
deliver one-stop services, including content delivery, cloud storage, big data, and
video cloud services. In addition, CDN vendors provide 24/7 O&M and monitoring
to ensure network connectivity at any time.
• Huawei Cloud CDN caches origin content on edge nodes across the globe. When
a user accesses the content, the user does not need to retrieve it from the origin
server. Based on a group of preset policies (including content types, geological
locations, and network loads), CDN provides the user with the IP address of a
CDN node that responds the fastest, enabling the user to obtain the requested
content faster than would have otherwise been possible.
• Huawei Cloud CDN has over 2,000 edge nodes in the Chinese mainland and over
800 edge nodes outside the Chinese mainland. The network-wide bandwidth is at
least 150 Tbit/s. Edge nodes are deployed on networks of top carriers in China
such as China Telecom, China Unicom, China Mobile, and China Education and
Research Network (CERNET), as well as many small- and medium-sized carriers.
Up to now, Huawei Cloud CDN covers more than 130 countries and regions,
connecting to over 1,600 carrier networks. CDN precisely schedules user requests
to the most appropriate node for efficient and reliable acceleration.
• Dynamic data: web program
▫ Cloud server backup. This type of backup uses the consistency snapshot
technology for disks to protect data of ECSs and BMSs. The backups of
servers without deployed databases are common server backups, and those
of servers with deployed databases are application-consistent backups.
▫ SFS Turbo backup. This type of backup protects data of SFS Turbo file
systems.
• AB
• Discussion 1:
• Discussion 2:
• Typical products:
• Huawei Cloud relational database services include RDS for MySQL, RDS for
PostgreSQL, RDS for SQL Server, GaussDB(for openGauss), and GaussDB(for
MySQL).
• Access control
• Transmission encryption
▫ You can download the Certificate Agency (CA) certificate from the console
and upload it when connecting to a database for authentication.
• Storage encryption
▫ RDS encrypts data before storing it. Encryption keys are managed by Key
Management Service (KMS).
• Data deletion
▫ Automated backup data and the data stored in the disks associated with
your instance can be securely deleted. You can restore a deleted DB
instance from a manual backup or rebuild the DB instance in the recycle bin
during the retention period.
• RDS for MySQL
▫ It uses a stable architecture and supports a wide range of web applications.
It is cost-effective and often preferred by small and medium enterprises.
▫ A web-based console is available for you to monitor performance metrics
so if there is an issue, you can identify it and take appropriate measures as
soon as possible.
▫ You can flexibly scale resources to meet business needs and pay for only
what you use.
• RDS for PostgreSQL
▫ RDS for PostgreSQL supports the postgis plugin and provides excellent
spatial performance.
▫ RDS for PostgreSQL is a cost-effective solution suitable for many business
scenarios. You can flexibly scale resources based on your needs and pay for
only what you use.
• RDS for SQL Server
▫ RDS for SQL Server is reliable, scalable, inexpensive, and easy to manage. It
supports high availability for your applications with automatic database
failover that completes within several seconds. It also provides multiple
options for backing up your data.
• Database engine versions: MySQL 5.6, 5.7, and 8.0
• Data security: Multiple security policies protect databases and data privacy.
• Database reliability: Three-copy data storage ensures up to 9 nines of database
data reliability and up to 11 nines of backup data reliability.
• High availability (intra-city disaster recovery): Primary/standby DB instances are
deployed within an AZ or across AZs, ensuring service availability over 99.95%.
• Instance access: Multiple access methods are supported. You can use floating IP
addresses, public IP addresses, or VPNs.
• Instance management: You can add, delete, modify, query, and reboot your DB
instance on the console.
• Elastic scaling: Horizontal scaling: Read replicas can be created (up to five for
each instance) or deleted. Vertical scaling: DB instance classes can be modified
and storage space can be scaled up to 10 TB.
• Backup and restoration:
▫ For backup, there are automated backup, manual backup, full backup, and
incremental backup. Backups can be added, deleted, queried, or replicated.
▫ For restoration, data can be restored to any point in time within the backup
retention period, or to a new or an original DB instance. The backup
retention period is up to 732 days.
• When creating a DB instance, you can select Primary/Standby as the instance
type. If a primary instance fails, RDS automatically switches to the standby
instance. If the standby instance also fails, a primary/standby instance in another
AZ will automatically take over the workloads.
• Each RDS DB instance supports up to five read replicas and can scale out with
Distributed Database Middleware (DDM) to further increase capacity. Write
requests are routed to the primary instance and read requests are routed to read
replicas.
• The primary and standby DB instances share the same virtual IP address (VIP) for
communication with external systems. The DB instance associated with the VIP is
the primary instance. If the primary instance is unavailable, RDS automatically
associates the VIP with the standby instance and promotes it to be the new
primary instance. Associating the VIP with the standby instance can be completed
in seconds. There is no downtime. The switchover is imperceptible to users.
• Constraints: You can create read replicas only after purchasing a DB instance.
• After read replicas are created and read/write splitting is enabled for your DB
instance, RDS will distinguish between read and write requests. Write requests
are routed to the primary instance. Read requests are distributed to the read
replicas.
• The automated backup retention period (1-732) is configurable.
• DB engine versions: 9.5, 9.6, 10.0, 11, and 12
• Data migration: There is online and offline migration to the cloud, to on-
premises, and across clouds.
• Horizontal scaling: Read replicas (up to five for each instance) can be created or
deleted. Vertical scaling: DB instance classes can be modified and storage space
can be scaled without downtime.
• Backup and recovery: RDS supports automated and manual backups along with
point-in-time recovery (PITR).
• RDS for PostgreSQL supports cross-AZ HA. If the primary instance fails, the fault
detection module attempts to start it three times. If the primary instance still
cannot be started, a failover is automatically performed and completed within
seconds. The standby instance is promoted to primary and read replicas are
automatically associated with the new primary instance.
• RDS provides data backup and restoration. You can set an automated backup
policy to back up your data daily. Automated backups can be retained for up to
732 days. An incremental backup is performed every 5 minutes for data
consistency.
• If data is lost or deleted by mistake, you can restore the database to any point in
time.
• Backup files are stored in OBS. OBS has no capacity upper limit and provides
99.999999999% data reliability.
• Note: Both of PostgreSQL and MySQL can be used in most scenarios.
▫ When you are choosing a database, database use and design habits need to
be considered. For example, some gaming and Internet companies just use
databases to store data. Both PostgreSQL and MySQL are fine for this. But
if many of your system's functions depend on more varied database
features, PostgreSQL is recommended. It is a stable and reliable open-
source database that is a good choice for many companies.
▫ If your current DB system only processes transactions, choose a database
using the same engine. If your database requires both transaction and
analytic processing, PostgreSQL is recommended because it provides
excellent analytical performance.
▫ If many stored procedures are used, PostgreSQL is recommended. Use
whatever your company is already used.
▫ If your application has to access heterogeneous databases, PostgreSQL is
recommended because it provides foreign data wrappers, which allows
users to access heterogeneous data using SQL statements.
▫ PostgreSQL is recommended for complex data types, such as complex
arrays, spatial data, network data, JSON data, XML data, and certain
custom types.
• PostgreSQL is recommended if your application requires geographic, spatial,
image, time series, multi-dimensional data, access to heterogeneous DB, machine
learning, text retrieval, or word segmentation and you do not want another
dedicated database.
• Shared DFV storage:
• Active-active architecture:
▫ GaussDB(for MySQL) does not support standby instances. All read replicas
are active, offloading read traffic from the primary node and improving
resource utilization.
▫ GaussDB(for MySQL) does not use page flushing or double writes. All
update operations are recorded in logs to save bandwidth.
• In TPC-H testing, if a DB instance (with 32 vCPUs and 256 GB of memory)
handles 100 GB of data, its performance is improved by 8x when handling of 16-
thread concurrency requests.
• Linear expansion of GaussDB(for MySQL) read and write performance:
• You do not need to re-divide storage for the new nodes because GaussDB(for
MySQL) uses DFV distributed storage. The new nodes can share the same storage
as the existing nodes.
• When data is restored, GaussDB(for MySQL) can provide services before the
restoration is complete. In contrast, traditional databases need to wait for all
data to be fully restored before they can provide services again.
• High security:
• Comprehensive tools
• Open-source ecosystem
• Player information, such as player items and bonus points, is stored in DDS
databases. During peak hours, DDS cluster instances can handle large amounts
of concurrent requests. DDS clusters and replica sets provide high availability to
ensure games are stable in high-concurrency scenarios.
▫ Ultra-high write performance, making GaussDB NoSQL a huge fit for IoT
and financial fraud detection scenarios
• High cost-effectiveness
▫ All data is stored in disks with cold and hot data separated. Hot data can
be read from the cache directly, making programs run fast.
• Hitless scaling
▫ Hot data is loaded to the memory and cold data is stored persistently, so
there is no need to use an extra MySQL database.
▫ Cold and hot data is automatically exchanged, making coding easier than
before.
GaussDB(for Redis) is Redis-compatible and can store a large amount of data
inexpensively and reliably, so it is a great fit for persistent storage scenarios.
Gaming:
• In most cases, you can migrate databases using both UGO and DRS. When
migrating databases from on-premises or other clouds to Huawei Cloud, you can
use UGO to analyze the source databases and migrate the databases based on
the actual scenario and the suggestions provided by UGO. You can also use the
full + incremental migration provided by DRS to migrate data from one database
to another.
• Easy to use
• Fast setup
▫ Traditional migration takes several days, weeks, or even months to set up.
• Low costs
• Secure
▫ The cost of labor for a typical database migration is used as a baseline, and
then the workloads involved in automatic database migration are added in.
Additionally, UGO evaluates the migration workloads based on the amount
of code involved, the conversion rate, and how hard it will be to modify
incompatible objects.
• Core Features 5: Database schema migration
▫ After evaluating the source database, UGO allows users to filter the objects
to be migrated, and then verifies and migrates the objects. Failed objects
are modified and the process is repeated until all objects pass.
• ABC
• ABCD
• Discussion 1:
• Discussion 2:
▫ Costs: Evaluate the costs and performance of different engines. When the
number of access requests is small, reduce the cluster scale.
• Demilitarized Zone (DMZ) is a special network area different from the external
network or internal network. Generally, the DMZ houses public servers that do
not contain confidential information, such as web servers, email servers, or FTP
servers. Users from the external network can only access the services in the DMZ,
but cannot access the information on the internal network. So, the information
on the internal network cannot be impacted even if the servers in the DMZ were
attacked.
• In an SQL injection attack, an attacker tricks the database server into executing
unauthorized queries. Attackers use exploits or logic flaws in application code to
bypass security controls. They manipulate the database server behind a web
application, tricking the system into doing what they want by executing specially
constructed SQL statements.
• Cross-Site Scripting (XSS) is a common type of web security vulnerability.
Attackers can exploit XSS vulnerabilities to inject malicious scripts into web pages
that are provided for other users. In most types of attacks, there are only two
parties involved: the attacker and the site they attack, but in an XSS attack, web
clients, and web applications are also involved, so website visitors also suffer. XSS
attacks are designed to steal cookies stored on a client or sensitive information
used by other websites to identify a client.
• In command injection attacks, attackers construct and submit special command
strings to embedded or web applications as these applications typically do not
check data submitted by users very strictly. After receiving the constructed
commands, applications are tricked into executing external programs or
launching OS attacks so that attackers can steal data or network resources.
• In a Trojan attack, attackers upload a Trojan to a legitimate website. When a
user visits the website, the Trojan is downloaded and executed automatically. The
user's computer is attacked and even manipulated by the attacker.
• Challenge Collapsar (CC) attacks are web attacks against web servers or
applications. In CC attacks, attackers send a large amount of standard GET/POST
requests to target system to exhaust web servers or applications. For example,
attackers can send requests to URIs of databases or other resources to make the
servers unable to respond to normal requests.
• A zero-day vulnerability is a vulnerability in a system or device that has been
disclosed but has not been patched yet. No one except the one who discovered
the vulnerability is aware of it. This person may exploit the vulnerability to
launch attacks, and such attacks are often unpredictable and destructive.
• The solid line indicates the access traffic.
• Demilitarized Zone (DMZ) is a special network area different from the external
network or internal network. Generally, the DMZ houses the public servers that
do not contain confidential information, for instance, web servers, email servers,
or FTP servers. Users from the external network can only access the services in
the DMZ, but cannot access the information on the internal network. So, the
information on the internal network cannot be impacted even if the servers in
the DMZ were attacked.
The global average total cost of data breaches increased by 10% from 2020 to
2021.
• Application scenarios:
▫ Data masking: The DSC data masking engine leverages a wide range of
preset and user-defined masking algorithms. It then masks structured and
unstructured data for storage.
• KMS is used for cloud service encryption (integrated in cloud services), data disk
encryption, and small-size data encryption.
• Private keys and passwords are not statically stored on the user side, reducing
the risk of private key and password leakage. KMS and KPS manage private keys
and regularly rotate keys in a unified manner, reducing the attack time window.
Private keys and passwords are encrypted by KMS/KPS on the cloud and then
securely stored. They are dynamically obtained after IAM/MFA authentication.
They are easy to use and can be accessed anytime, anywhere. Users can use IAM
credentials and MFA to obtain private keys and passwords anywhere to access
resources.
• Users or applications can use CSMS to create, retrieve, update, and delete
credentials in a unified manner throughout the credential lifecycle. CSMS can
help you eliminate risks that stem from insecure practices such as hardcoding,
plaintext configuration, and inadequate permission control.
• In this figure, DEW modules include KPS and KMS.
• Verizon is the largest wireless carrier in the United States, with over 140 million
subscribers.
• In digital transformation, companies face stringent security compliance
requirements. Complying with security requirements is a huge responsibility, and
non-compliance may result in severe penalties. Security compliance is the first
and most important thing that enterprises are concerned with in cloud migration.
Compliance standards determine the security level companies need to be able to
comply with on the cloud.
• A project can contain different resources. You can attach policies to different user
groups to grant permissions for accessing specific resources. In the figure, user A
is granted access to all resources in project A and to specific resources in project
B. User B is granted access to specific resources in project B and all resources in
project C.
• AK: An access key ID is a unique ID associated with an SK. An AK is used together
with an SK to cryptographically sign requests.
• You can assign permissions to IAM users through user groups. By default, new
IAM users do not have any permissions assigned. To assign permissions to new
users, add them to one or more groups, and assign permissions to these groups.
The users then inherit permissions from the groups they belong to, and they can
perform operations on cloud services based on the assigned permissions.
• Authorization policies:
• Cloud service delegation: Huawei Cloud services interwork with each other. Some
cloud services depend on other services. You can create an agency to delegate a
cloud service to call other services on your behalf. For example, if Container
Guard Service (CGS) needs to scan container images, you need to delegate
SoftWare Repository for Container (SWR) permissions to CGS.
• OpenID Connect (OIDC): a standard identity authentication protocol that runs on
top of the OAuth 2.0 protocol.
• Security Assertion Markup Language (SAML): Security Proposition Markup
Language. It is an XML-based open-standard for transferring identity data
between two parties: an identity provider (IdP) and a service provider (SP).
• Identity provider (IdP): collects and stores user identity information, such as
usernames and passwords, and authenticates users during login. For identity
federation between an enterprise and Huawei Cloud, the IdP refers to the identity
authentication system of the enterprise.
• Identity federation process:
▫ Create an IdP and establish a trust relationship.
▪ OIDC-based IdP: Create OAuth 2.0 credentials in the enterprise IdP
and create an IdP in Huawei Cloud to establish a trust relationship
between the enterprise and Huawei Cloud.
▪ SAML-based IdP: Exchange the metadata files (SAML 2.0-compliant
interface files that contain interface addresses and certificate
information) of the enterprise IdP and Huawei Cloud. Then, create an
IdP in Huawei Cloud to establish a trust relationship between the
enterprise and Huawei Cloud.
▫ Configure identity conversion rules: Map the users, user groups, and their
permissions in the enterprise IdP to Huawei Cloud.
▫ Configure a login link: Configure a login link in the enterprise management
system to allow users to access Huawei Cloud using SSO.
• After data is collected, it is batch processed by the big data platform and then
analyzed by the big data operations center. Analysis results are reported to SA so
that SA can take appropriate protective actions such as event analysis and alarm
reporting.
• Asset management: As enterprises migrate more workloads to the cloud, more
cloud assets are used, and there are frequent changes made to those assets. This
means more security risks on the cloud.
▫ SA gives customers a comprehensive view of the security status of assets on
the cloud. SA monitors the security status of all assets in the cloud in real
time and visualizes vulnerabilities, threats, and attacks on servers, making it
easier for customers to handle risks.
• Threat event alarms: Security threats to clouds never stop, and a variety of new
threats are emerging every day.
▫ By collecting network-wide traffic data and security device logs, SA can
detect and monitor security risks on the cloud in real time, display statistics
on security events in real time, and aggregate event data from other
security services. SA uses preset security policies to effectively defend
against common brute-force attacks, web attacks, Trojans, and zombie
bots, greatly improving defense and O&M efficiency.
• Vulnerability notifications: Service security is of top priority during cloud
migrations. To prevent vulnerabilities from being exploited, we need to find and
fix as many vulnerabilities as possible.
▫ Apart from reporting latest vulnerabilities based on emergency security
notices issued on Huawei Cloud, SA periodically scans OSs, software, and
websites for vulnerabilities by working with linked security services, making
it easier for customers to centrally manage server and website
vulnerabilities. SA also provides mitigation suggestions. With centralized
vulnerability management on the cloud, SA helps customers quickly identify
key risks and vulnerable assets and harden their service system.
• SA can scan for unsafe settings of cloud services, report scan results by category,
generate alarms for unsafe settings, and provide hardening suggestions and
guidelines.
• MTD collects logs from IAM, DNS, CTS, OBS, and VPC and uses an AI engine,
threat intelligence, and detection policies to continuously detect potential threats,
malicious activities, and unauthorized behavior, such as brute-force cracking,
penetration attacks, and mining attacks.
• MTD identifies threats to IAM accounts and vulnerabilities to DNS and looks for
intrusions by checking CTS logs. These security risks cannot or can barely be
detected by other security services. When risks increase, multi-factor verification
or biometric recognition is required by MTD for using an IAM account.
• 1. Answer: False.
• 2. Answer: A.
• Discussion 2:
▪ FaaS and BaaS products can be scaled flexibly and precisely to process
each request. For developers, a serverless platform does not need
capacity planning or auto scaling triggers or rules.
▫ Update resources accordingly. After the migration, cluster resources may fail
to be deployed. You need to update the faulty resources. The possible
adaptation problems lie in images, Services and ingresses, StorageClasses,
and databases.
▫ Perform additional tasks. After cluster resources are properly deployed,
verify application functions after the migration and switch service traffic to
the target cluster. After confirming that all services are running properly,
bring the source cluster offline.
• CCE is deeply integrated with high-performance HUAWEI CLOUD computing
(ECS/BMS), network (VPC/EIP/ELB), and storage (EVS/OBS/SFS) services, and
supports heterogeneous computing architectures such as GPU and Arm. You can
build high-availability Kubernetes clusters secured by multi-AZ, cross-region
disaster recovery (DR) and auto scaling.
• Huawei is amongst the first developers of the Kubernetes community in China.
Huawei is a major contributor to the open source community and a leader in the
container ecosystem. Huawei Cloud CCE is the earliest commercial Kubernetes
service in China, and is also one of the first products that passed the CNCF
Certified Kubernetes Conformance Program. CCE features benefits such as access
to open ecosystems, enhanced commercial features, and adaptation to
heterogeneous infrastructure.
• Volcano: Native Kubernetes has weak support for batch computing services.
Volcano provides two enhanced batch computing capabilities. One is advanced
job management, such as task queuing, priority setting, eviction, backfilling, and
starvation prevention. The other is intelligent scheduling, such as topology-aware
affinity-based scheduling and dynamic driver-executor ratio adjustment. In
addition, scheduling and distributed frameworks such as gang scheduling and PS-
Worker are supported.
• You can use CCE via the CCE console, kubectl, or Kubernetes APIs.
• A node is a basic element of a container cluster. CCE uses high-performance
Elastic Cloud Servers (ECSs) or Bare Metal Servers (BMSs) as nodes to build
highly available Kubernetes clusters.
• Kata containers are distinguished from common containers in a few aspects. The
most important difference is that each Kata container (pod) runs on an
independent micro-VM, has an independent OS kernel, and is securely isolated at
the virtualization layer. CCE provides container isolation that is more secure than
independent private Kubernetes clusters. With Kata containers, kernels,
computing resources, and networks are isolated between different containers to
protect pod resources and data from being preempted and stolen by other pods.
• A workload is an application running on Kubernetes. No matter how many
components are there in your workload, you can run it in a group of Kubernetes
pods.
• CCE supports Kubernetes-native deployment and lifecycle management of
container workloads, including creation, configuration, monitoring, auto scaling,
upgrade, uninstall, service discovery, and load balancing.
• Recommendations on CIDR block planning:
▫ CIDR blocks cannot overlap. Otherwise, a conflict occurs. All subnets
(including those created from the secondary CIDR block) in the VPC where
the cluster resides cannot conflict with the container and Service CIDR blocks.
▫ Ensure that each CIDR block has sufficient IP addresses. The IP addresses in
the node CIDR block must match the cluster scale. Otherwise, nodes cannot
be created due to insufficient IP addresses. The IP addresses in the container
CIDR block must match the service scale. Otherwise, pods cannot be created
due to insufficient IP addresses.
• In the Cloud Native Network 2.0 model, the container CIDR block and node CIDR
block share the IP addresses in the same VPC. Therefore, you are advised not to
set the container subnet and node subnet to the same. Otherwise, containers or
nodes may fail to be created due to insufficient IP resources.
• CCE supports the following container network models: container tunnel network,
VPC network, and Cloud Native Network 2.0.
• The container tunnel network is constructed on but independent of the node
network through tunnel encapsulation. This network model uses VXLAN to
encapsulate Ethernet packets into UDP packets and transmits them in tunnels.
Open vSwitch serves as the backend virtual switch. Though at some costs of
performance, packet encapsulation and tunnel transmission enable higher
interoperability and compatibility for most scenarios that do not require high
performance.
• Advantages: The container network directly uses the VPC, making it easy to locate
network problems and improve the networking performance. Requests from
external networks in a VPC can be directly routed to a container IP address. Load
balancing, security groups, and EIPs provided by the VPC can be directly used.
• Disadvantages: The container network consumes the IP addresses in the VPC. You
need to plan the container CIDR block before creating a cluster.
• This network model is available only to CCE Turbo clusters.
• In CCE, container storage is backed both by Kubernetes-native objects, such as
emptyDir, hostPath, secret, and ConfigMap, and by cloud storage services. These
cloud storage services can be accessed via Container Storage Interface (CSI).
• CSI enables Kubernetes to support various classes of storage. For example, CCE
can easily interconnect with Huawei Cloud block storage (EVS), file storage (SFS),
and object storage (OBS).
• CCE provides an add-on named everest to serve as CSI. Everest is a cloud native
container storage system. Based on CSI, clusters can interconnect with Huawei
Cloud storage services such as EVS, OBS, SFS, and SFS Turbo. everest is a system
resource add-on. It is installed by default when a cluster of Kubernetes v1.15 or
later is created.
• Ease of use:
▫ You can directly push and pull container images without platform build or
O&M.
▫ SWR provides an easy-to-use management console for full lifecycle
management over container images.
• Security and reliability:
▫ SWR supports HTTPS to ensure secure image transmission, and provides
multiple security isolation mechanisms between and inside accounts.
▫ SWR leverages professional storage services of Huawei to ensure reliable
image storage.
• Faster image pull and build:
▫ P2P acceleration technology developed by Huawei brings faster image pull
for CCE clusters during high concurrency.
▫ Intelligent node scheduling around the globe ensures that your image build
tasks can be automatically assigned to the idle nodes nearest to the image
repository.
• From the practices of customers and partners, there are four typical scenarios of
using CCE:
▫ First, progressive IT architecture upgrade. With CCE, complex applications in
traditional architectures are decoupled into multiple lightweight modules.
Each module is run as a Kubernetes workload. For example, stateless
applications run as Deployments and stateful applications run as
StatefulSets. In this way, modules can be flexibly upgraded and scaled to
meet changing market demands.
▫ Second, faster service rollout. The same container image can be used
through each phase from R&D to O&M to ensure the consistency of service
running environments. Services can be used out of the box and rolled out
faster.
▫ Third, auto scaling upon service traffic fluctuation. Containers can be quickly
scaled within seconds to ensure service performance.
▫ Fourth, fewer resources and reduced cost. With containers, host resources
can be divided at a finer granularity to improve resource utilization.
• In the serverless model, a cloud provider runs servers and dynamically allocates
resources so that you can build and run applications without having to create,
manage, or maintain servers. This model helps you improve development
efficiency and reduce IT costs.
• CCE provides semi-hosted clusters, while CCI provides fully-hosted clusters that do
not need manual management.
• Functions:
▫ CCI provides one-stop container lifecycle management, allowing you to run
containers without creating or managing server clusters.
▫ CCI supports multiple types of compute resources, including CPUs, GPUs, and
Ascend chips, to run containers.
▫ Various network access modes and layer-4 and layer-7 load balancing are
available to meet scenario-specific needs.
▫ CCI can store data on various Huawei Cloud storage volumes, including EVS,
SFS, and OBS.
▫ CCI supports fast auto scaling. Users can customize scaling policies and
combine multiple scaling policies to cope with traffic surge during peak
hours.
▫ The comprehensive container status monitoring of CCI monitors the
resources consumed by containers, including the CPU, memory, GPU, and
GPU memory usage.
▫ CCI provides dedicated container instances, which run Kata containers on
high-performance physical servers, enabling VM-level security isolation
without performance deterioration.
• With CCI, you can stay focused on your own services, instead of underlying
hardware and resources. CCI is billed by the second for convenient use anytime.
• Dedicated container instances allow you to exclusively use physical servers and
support service isolation among departments. They run Kata Containers on high-
performance physical servers, enabling VM-level security isolation without
performance loss. Huawei Cloud performs O&M, allowing you to completely
focus on your services.
• CCI provides VM-level isolation without compromising the startup speed, offering
you better container experience. It has the following features:
▫ Native support for Kata containers
▫ Kata-based kernel virtualization, providing comprehensive security isolation
and protection
▫ Huawei-developed virtualization acceleration technologies for higher
performance and security
• Currently, most big data and AI training and inference applications (such as
TensorFlow and Caffe) run in containers. These applications are GPU intensive
and require high-performance network and storage. As these applications are
task-based, resources must be quickly allocated upon task creation and released
upon task completion, and powerful compute and network resources as well as
high I/O storage are required for high-density computing.
• CCI resources are billed on demand by second, reducing costs.
• Volcano is a batch processing platform based on Kubernetes. It provides a series
of features required by machine learning, deep learning, bioinformatics, genomics,
and other big data applications, as a powerful supplement to Kubernetes
capabilities. Volcano provides general-purpose, high-performance computing
capabilities, such as job scheduling, heterogeneous chip management, and job
running management, serving end users through computing frameworks for
different industries, such as AI, big data, gene sequencing, and rendering.
(Volcano has been open-sourced in GitHub.)
• No O&M is required for clusters and servers, which greatly reduces costs.
• CCI is tailored for task-based scenarios.
▫ These scenarios include heterogeneous hardware-based AI training and
inference, training tasks can be hosted on CCI.
▫ It also works in HPC scenarios, such as gene sequencing.
▫ Third, burst scale-out in a long-term stable running environment, such as e-
commerce flash sales and hot topic-based marketing.
• The main advantages of CCI are on-demand use for lower costs, and full hosting
for O&M-free. It also enables consistency and scalability based on standard
images.
• CCI supports pay-per-use or package-based billing. A core-hour indicates the
number of cores multiplied by time. For example, 730 core-hours indicate that
you can use 730 cores for one hour or one core for 730 hours.
▫ In pay-per-use mode, you will be charged by second for each instance and
the billing statistics are presented by hour.
▫ In package-based billing mode, if your resource usage exceeds the quota of
the package within the package validity period, you will be billed for the
excess usage on a pay-per-use basis. If you buy multiple packages, resources
in the package with the earliest expiration time will be used first.
• To work with AOS, you only need to create a template describing the applications
and the required cloud resources, including their dependencies and references.
AOS will then set up these applications and provision the resources as specified in
the template. For example, when creating an ECS, together with a VPC and a
subnet on which the ECS runs, you only need to create a template defining an
ECS, VPC, subnet, and their dependencies. AOS will then create a stack, namely, a
collection of resources you specified in the template. After the stack has been
successfully created, the ECS, VPC, and subnet are available to use.
• Product functions:
▫ AOS provides automatic orchestration of mainstream Huawei Cloud services.
For details, see Cloud Services and Resources that Can Be Orchestrated in
AOS. AOS also provides lifecycle management including resource scheduling,
application design, deployment, and modification, to reduce O&M costs
through automation.
▫ Standard languages (YAML and JSON) can be used to describe required
basic resources, application systems, upper-layer services, and their
relationships. Automatic resource provision, application deployment, and
service loading can be implemented in a few clicks based on uniform
description and defined dependency relationships. You can manage deployed
resources and applications in a unified manner.
▫ AOS Template Market provides abundant templates for free, including basic
resource templates, service combination templates, and industry templates,
covering common application scenarios. You can use public templates
directly to deploy services in the cloud in a few clicks.
• Karmada is a multi-cluster management system built on Kubernetes native APIs.
It provides automated multi-cluster management capabilities in a pluggable
manner for multi-cloud and hybrid cloud applications. Karmada enables
centralized management, high availability, fault recovery, and traffic scheduling.
• MCP leverages cluster federation to implement unified management of clusters
of different cloud service providers. As a unified entry for multiple clusters, MCP
supports dynamic cluster access and global cluster monitoring dashboard.
• Based on the multi-cluster and federation technologies, MCP manages
Kubernetes clusters across regions or clouds and supports full lifecycle
management of applications across clusters, including deployment, deletion, and
upgrade, by using standard cluster federation APIs in Kubernetes.
• MCP supports cross-cluster auto scaling policies to balance the pod distribution in
each cluster and implement global load balancing.
• You can create federated Services for cross-cluster service discovery. MCP enables
service region affinity based on the proximity access principle, reducing network
latency.
• MCP is compatible with the latest Kubernetes-community federation architectures,
Kubernetes native APIs and Karmada APIs.
• MCP supports application federation, which allows you to deploy an application
from only one cluster to multiple clusters across clouds in just a few clicks. In this
way, cross-cloud DR and traffic sharing are implemented.
• You can clone or migrate your applications to other clusters or across
clouds/regions in just a few clicks without re-writing or modifying your service
code.
• Service release: Service providers upload a service package, verify the lifecycle and
features of the service in the OSC, and release the service as an offering for other
tenants to subscribe to.
• Service subscription: OSC contains Huawei-developed services, services published
by ecosystem partners, and open source services. All services can be subscribed to
by users. Instances can be deployed only after successful subscription.
• Service unsubscription: Users can unsubscribe from a service at any time. Upon
unsubscription, the system automatically deletes the deployed services and
instances.
• Private service uploading: Users can upload services developed based on Helm,
Operator Framework, or OSC service specifications to OSC as private services for
management.
• Service upgrade: When a provider publishes the updated version of a service, the
subscribers will receive an upgrade notification and can decide whether to
upgrade the service to the latest version.
• Instance deployment: After subscribing to a service, users can deploy an instance,
specifying the region, container cluster, and running parameters.
• Instance O&M: OSC provides the O&M view of instances. Users can view the
monitoring and logs of instances and switch from the O&M view to the
corresponding cloud service for in-depth data analysis.
• Instance update: Users can modify the running configurations of an instance.
• Instance deletion: When the lifecycle of a service running in an instance ends,
users can delete the instance to reclaim related resources.
• Serverless computing does not mean that we no longer use servers to host and
run code, nor does it mean that O&M engineers are no longer needed.
Conversely, it means that consumers no longer need to spend time and resources
on configuring, maintaining, updating, or expanding servers, or planning capacity.
All these are handled by a serverless platform, enabling developers to focus on
service logic and O&M engineers to process key service tasks.
• There are two serverless architectures:
▫ Functions-as-a-service (FaaS): provides event-driven computing. Developers
use functions triggered by events or HTTP requests to run and manage
application code. They deploy small units of code to FaaS, where the code is
executed as discrete actions on request, and can be expanded without
managing servers or any other underlying infrastructure.
▫ Backend-as-a-service (BaaS): an API-based third-party service that can
replace the core function subset in applications. Because these APIs are
provided as services that can be automatically expanded and transparently
operated, they are serverless for developers.
• FaaS executes function code, and BaaS only uses APIs to provide backend services
on which applications depend.
• Generally, serverless is recommended for workloads in the following
scenarios:Asynchronous, concurrent, easy to be parallelized into independent
units.Infrequent requests with huge and unpredictable expansion
requirements.Stateless and transient, without instant cold start
requirements.Highly dynamic service requirement changes.
• Serverless products or platforms have the following benefits:
• No server O&M: Serverless has significantly changed the application cost model
by eliminating the overhead involved in maintaining server resources.
▫ No need to configure, update, or manage servers. Managing servers, VMs,
and containers involves personnel, tools, training, and time.
▫ FaaS and BaaS products can be scaled flexibly and precisely to process each
request. For developers, a serverless platform does not need capacity
planning or auto scaling triggers or rules.
• No cost for idle resources: For consumers, a major benefit of serverless products is
that idle resources do not incur any cost. For example, idle VMs and containers
will not be charged. However, the costs for stateful storage, functions, and
feature sets will be charged.
• When using FunctionGraph, you do not need to apply for or pre-configure any
compute, storage, or network services. You only need to upload and run code in
supported runtimes. FunctionGraph provides and manages underlying compute
resources, including CPUs, memory, and networks. It also supports configuration
and resource maintenance, code deployment, automatic scaling, load balancing,
secure upgrade, and resource monitoring.
• FunctionGraph supports Node.js, Java, Python, Go, and C#, allowing you to edit
code inline, import OBS files, and upload ZIP and JAR packages. It uses SMN,
APIG, and OBS triggers. It collects and displays real-time metrics and logs, and
enables you to query logs online, making it easy to view function status and
locate problems. Function flows orchestrate and coordinate multiple distributed
functions. FunctionGraph provides unified plug-ins for on-/off-cloud development
and debugging. HTTP functions can be triggered for web service optimization by
sending HTTP requests to specific URLs. In addition, you can enable tracing on the
function configuration page so that you can view Java virtual machine (JVM) and
tracing information on the APM console. Currently, this feature is only available
for Java functions. You can package and upload container images to
FunctionGraph for running.
• FunctionGraph 2.0 is a next-generation function computing and orchestration
service. It has the following features:
▫ Deep integration with CloudIDE, concurrent function debugging, tracing,
wizard-based building, and full lifecycle management
▫ Six programming languages and custom runtime, cold startup and auto
scaling in 100 milliseconds
▫ First to support stateful functions in China, visualized function orchestration
▫ Serverless web applications with zero reconstruction
• Application development: out-of-the-box CloudIDE, debugging and tracing of
clustered serverless applications, code breakpoints, stack viewing, call topologies,
and hot code replace (HCR)
• CI/CD: deep integration with serverless runtimes; lightweight DevOps with O&M
tools
• Application hosting: lifecycle management with unified specifications; templates
and marketplace for experience and reuse
• Cloud application engine (CAE): a one-stop serverless application hosting service
that enables ultra-fast deployment at low cost with simple O&M. It releases
applications from source code, software packages, and image packages, with
seconds of auto scaling, pay-per-use billing, no infrastructure O&M, and multiple
observable metrics.
• (On-cloud) CloudIDE: Create a function using a template, view the function and
download it to the cloud, debug it using CloudIDE, and push it to the cloud.
• (Off-cloud) VSCode plug-in: Create a function using a template, view the function
on the cloud, download it to a local host, debug it using VSCode plug-in, and
push it to the cloud.
• HTTP functions are better for optimizing web services and can be triggered by
sending HTTP requests to specific URLs. You can specify this type when creating a
function. HTTP functions only support APIG and APIC triggers.
• The following challenges may exist when you shift from the traditional
development mode to the serverless mode:
▫ Different runtimes and deliverable formats: The runtime provided by
serverless function vendors may be Docker or microVM. The deliverable
formats and function signatures are different. You have to make adaptations.
▫ Immature ecosystem: Popular open-source tools (such as CI/CD pipeline) are
not supported.
• The container ecosystem is mature and does not have portability and agile
delivery issues. Container images are standard deliverables in the cloud native
era. However, containers still involve O&M and idle resource costs.
• You can create custom images for both event and HTTP functions.
• Answer 1: False
▫ Label instead of ConfigMaps
• Answer 2: ABC
▫ Supports multiple languages, such as Node.js, Java, Python, Go, and C#
• Discussion 1:
▫ Construction cost, including equipment, site, and prices
▫ O&M costs, including manpower, power, and network costs
▫ Security
▫ Convenience
• Discussion 2:
▫ Response speed
▫ Performance
▫ Security
▫ Maintainability
▫ Cost
▫ Convenience
• Enterprises need to make a trade-off between rapid service development and
exquisite application architecture. Microservice architecture is the future trend. The
microservice architecture has abundant features, including fault tolerance, quick
rollout, more complex functions, high availability, requirement response,
manageability, and independent module release.
• On the monolithic architecture, all functions are integrated in one project. The
architecture is simple, the development cost in the early phase is low, and the
development period is short. Therefore, this architecture is ideal for small-scale
projects. However, as the small projects grow larger, it is difficult to develop,
expand, and maintain the monolithic architecture.
• Projects using the monolithic architecture are vertically divided, so small projects
cannot become too large.
• On the SOA, repeated common functions are extracted as components to provide
services for each system. Projects (or systems) communicate with services through
WebService or remote procedure call (RPC). SOA improves the development
efficiency and system reusability and maintainability.
• But the SOA has disadvantages. The boundary between systems and services is
blurred, which is not conducive to development and maintenance. The granularity
of the extracted services is too large, and systems are highly coupled with the
services.
• The microservice architecture is an approach to developing a single application as
a suite of small services, each running in its own process, and communicating with
lightweight mechanisms, often an HTTP resource API. Services are split at a finer
granularity, which facilitates resource reuse and improves development efficiency.
In this way, optimization solutions for each service can be formulated more
accurately, improving the system maintainability.
• The monolithic architecture is an archive package. The package contains
applications with all functions. In the early stage of software development, the
monolithic architecture is popular because it is easy to deploy, the technologies
are simple, and the labor cost is low. However, in the Internet era, the complexity
of service requirements and the delivery frequency increase. The traditional
monolithic architecture cannot meet the requirements of developers:
▫ The monolithic architecture is complex as all modules are coupled. They
have blurred boundaries and complex dependencies. Function adjustment
may bring unknown impacts and potential bugs.
▫ When a monolithic system encounters a performance bottleneck, the
system can only scale out horizontally and add service instances to balance
the load. Vertical expansion and module decoupling are not supported.
▫ The monolithic architecture has poor scalability. A monolithic application
can be scaled only as a whole. Scaling of a single module cannot be
performed.
▫ The monolithic architecture cannot isolate faults. The entire system may
break down even when a small module is faulty (for example, a request is
blocked) as all function modules are aggregated.
▫ On the monolithic architecture, the release impact is large. The entire
system is released each time and the system restarts upon each release.
This poses a great challenge to a large-scale integrated system. If we
decouple each module, only the modified module needs to be released.
▫ The deployment slows down. The build and deployment duration increases
as the code size increases.
▫ Technological innovation is hindered. A monolithic application solves all
problems using a unified technical platform or solution. Each team member
must use the same development language and architecture.
• SOA decouples applications, modularizes them, and builds functions into
independent units to provide services.
• SOA contains multiple services. The services communicate with each other
through mutual dependency or communication mechanisms to provide a series
of functions. A service independently exists in an OS process. Services invoke each
other through networks.
▫ First, system integration. SOA sorts out the mesh structure between
scattered and unplanned systems into a regular and governable star
structure. Some products, such as the ESB, technical specifications, and
service management specifications, need to be introduced.
▫ Second, system as a service. SOA abstracts service logic into reusable and
assemblable services and orchestrates the services to quickly regenerate
services. This transforms inherent functions into common services to quickly
reuse business logic.
• The microservice architecture provides a modular solution for functions that are
impossible in monolithic encoding. A single service is easy to develop, understand,
and maintain.
• Microservices are independently implemented and deployed, that is, they run in
independent processes. Therefore, they can be independently monitored and
expanded.
• Earlier, the data plane SideCar proxy is unstable and traffic-intensive as Service
Mesh puts too many functions, including the inter-service communications and
related governance into it. As a solution to these problems, the second-
generation Service Mesh emerges and separates the configuration policy and
decision logic from the proxy servers to form an independent control plane.
• Istio has two components: the data plane and the control plane.
▫ The control plane takes your desired configuration, and its view of the
services, and dynamically programs the proxy servers, updating them as the
rules or environments change.
• Istio service mesh has two components: the data plane (Envoy) and the control
plane (Istiod).
▫ Envoy is a high-performance proxy developed in C++ to mediate all
inbound and outbound traffic for all services in the service mesh. Envoy
proxies are deployed as sidecars to services and are the only Istio
components that interact with data plane traffic. In addition to load
balancing, circuit breakers, and fault injection, Envoy also supports a
pluggable extension model built on WebAssembly (Wasm) that allows for
custom policy enforcement and telemetry generation for mesh traffic.
▫ Istiod is a control plane component that provides service discovery,
configuration, and certificate management. Istiod converts advanced rules
written in YAML into Envoy-specific configurations and propagates them to
the sidecars. Pilot abstracts platform-specific service discovery mechanisms
and synthesizes them into a standard format that sidecars can consume.
Citadel enables strong service-to-service and end-user authentication with
built-in identity and credential management. You can also use Istio's
authorization feature to control who can access your services.
• As a core component of Istio, Pilot manages and configures all sidecar proxies
deployed in a specific Istio service mesh. As a component responsible for
configuration management, Galley verifies the format and content of the
configuration information and provides it for the Pilot on the control plane.
Citadel consists of the CA server, security discovery server, and certificate key
controller.
• Core concepts of Istio:
▫ Data plane components are injected as non-intrusive sidecars into service
containers, with transparent traffic hijacking.
▫ Upper-level APIs are implemented based on Kubernetes CRDs, fully
declarative and standardized.
▫ The data plane and control plane communicate with each other through
standard protocols, allowing pub/sub messaging.
• Istio extends Kubernetes to establish a programmable, application-aware
network using the powerful Envoy service proxy. Working with both Kubernetes
and traditional workloads, Istio simplifies deployment with standard, universal
traffic management, telemetry, and security.
• Istio aims to achieve scalability and meet various deployment requirements. Istio
control plane runs on Kubernetes. In this way, applications deployed in a cluster
can be added to your mesh. In addition, the mesh can be extended to other
clusters, and even connected with VMs or other endpoints running outside
Kubernetes.
• To enable Istio, you only need to deploy a special sidecar proxy in the
environment and use the Istio control plane to configure and manage the proxy
to intercept all network communication between microservices. You can use Istio
to achieve:
▫ Automatic load balancing for HTTP, gRPC, WebSocket, and TCP traffic
▫ Fine-grained control of traffic behavior with rich routing rules, retries,
failovers, and fault injection
▫ A pluggable policy layer and configuration API supporting access control,
rate limits and quotas
▫ Automatic metrics, logs, and traces for all traffic within a cluster, including
cluster ingress and egress
▫ Secure service-to-service communication in a cluster with strong identity-
based authentication and authorization
• Google Remote Procedure Call (gRPC) is a high-performance open source RPC
software framework built on the HTTP 2.0 transport layer protocol. It provides an
API design method for managing and configuring network devices. gRPC
supports multiple programming languages, such as C, Java, Golong, and Python.
• ASM supports smooth access and unified governance of multiple applications,
such as containers, traditional microservices, and third-party services. It enables
hybrid management of cross-cluster traffic under various network conditions in
multi-cloud and hybrid cloud scenarios. Large-scale meshes are provided for
intelligent O&M and scaling to help you automatically and transparently manage
application access.
• ASM provides a high-performance, low-loss, lightweight, and multi-form mesh
data plane and supports uninstallation by pod and node, accelerating sidecar
forwarding. Flexible topology learning optimizes configurations and resources on
the mesh control plane.
• ASM can well resolve application network governance issues such as challenges
in cloud native application management, network connection, and security
management.
• ASM is deeply integrated with CCE to manage application traffic and lifecycle in a
non-intrusive manner. ASM enhances the full-stack capabilities of Huawei Cloud
container services with better usability, reliability, and visualization.
• Hybrid deployment: Unified governance of hybrid deployment of VM applications
and containerized applications
• Observability: Out-of-the-box usability and end-to-end intelligent monitoring,
logs, topologies, and tracing
• Unified service governance in the multi-cloud and hybrid cloud scenarios, unified
service governance of multiple infrastructure resources (multi-container
cluster/container-VM/VM-PM), and cross-cluster grayscale release, topology, and
tracing
• Protocol extension: Solution of integrating with microservice SDKs for Spring
Cloud
• Community and open source: No. 3 in the world by contribution to Istio
community; quick response to community version issues and requirements
• Grayscale release policies:
▫ Grayscale policies based on request content: You can set criteria based on
request content, such as header and cookie. Only requests meeting the
criteria will be distributed to the grayscale version.
▫ Grayscale policies based on traffic ratio: You can set specific ratio for the
traffic to be distributed to the grayscale version.
▫ Canary release: Guidance will be provided to help you perform canary
release on a service, including rolling out a grayscale version, observing the
running and traffic of the grayscale version, configuring grayscale release
policies, and diverging the traffic.
▫ Blue-green deployment: Guidance will be provided to help you perform
blue-green deployment on a service, including rolling out a grayscale
version, observing the running and traffic of the grayscale version, and
switching the traffic.
• An O&M-free hosting control plane is provided. Unified service governance,
grayscale release, security, and service running monitoring capabilities for
multiple clouds and clusters are supported. Unified service discovery and
management of multiple infrastructure resources such as containers and VMs are
provided.
• The meshes of multiple clusters share a set of root certificates. They distribute
keys and certificate pairs to service pods in the data plane, and periodically
change key certificates. Key certificates can be revoked as required. When a
service calls another service, the mesh data plane envoy performs two-way
authentication and channel encryption. These two services can come from two
different clusters. Transparent end-to-end two-way authentication across clusters
is supported.
• Load balancing, service routing, fault injection, outlier detection, and fault
tolerance policies can be intuitively configured using an application topology.
Microservice traffic management can be real-time, visualized, intelligent, and
automated, requiring no modifications on your applications.
▫ Routing rules based on weight, content, and TCP/IP implements flexible
grayscale release of applications.
▫ HTTP sticky session achieves service processing continuity.
▫ Rate limiting and outlier detection ensure stable and reliable links
between services.
▫ Network persistent connection management saves resources and improves
network throughput.
▫ Service security certification, authentication, and audit lay a solid
foundation for service security assurance.
• Load balancing, service routing, fault injection, outlier detection, and fault
tolerance policies can be intuitively configured using an application topology.
Microservice traffic management can be real-time, visualized, intelligent, and
automated, requiring no modifications on your applications.
▫ Routing rules based on weight, content, and TCP/IP implements flexible
grayscale release of applications.
▫ HTTP sticky session achieves service processing continuity.
▫ Rate limiting and outlier detection ensure stable and reliable links between
services.
▫ Network persistent connection management saves resources and improves
network throughput.
▫ Service security certification, authentication, and audit lay a solid
foundation for service security assurance.
• Requests can be distributed based on the request content (browsers or OSs).
• Requests can be distributed based on traffic ratio.
• Container-based infrastructure brings a series of new challenges. It is necessary
to evaluate and enhance the performance of API endpoints and identify potential
risks of infrastructure. ASM enables you to enhance API performance with no
code refactoring and service delay.
• In traditional iterations, a new service version is directly released to all users at a
time. This is risky, because once an online accident or a bug occurs, the impact on
users is great. It could take a long time to fix the issue. Sometimes, the version
has to be rolled back, which severely affects user experience. Grayscale release is
a smooth iteration mode for version upgrade. During the upgrade, some users
use the new version, while other users continue to use the old version. After the
new version is stable and ready, it gradually takes over all the live traffic.
• Main features:
▫ Ease of use: Instances created in minutes; out of the box with visual
operations and real-time monitoring
• Clients:
• Broker: receives and processes requests from clients and persists messages.
• Rich features: DMS for RabbitMQ supports Advanced Message Queuing Protocol
(AMQP) and a variety of messaging features such as message broadcast, delayed
delivery, and dead letter queues.
• Monitoring and alarm: RabbitMQ cluster metrics are monitored and reported,
including broker memory, CPU usage, and network flow. If an exception is
detected, an alarm will be triggered.
• You can also obtain and call open APIs from APIG to reduce your development
time and costs.
• By using APIG, you can monetize services while reducing R&D investment for
more business focus and higher operational efficiency. For example, enterprise A
has created a mobile number location lookup API in APIG and released it on
KooGallery. Enterprise B obtains and calls the API from KooGallery and pays for
the fee incurred. In this way, enterprise A monetizes its services and enterprise B
reduces its development time and costs, achieving shared success.
• Swagger is a standard, complete framework for generating, describing, invoking,
and visualizing RESTful web services. It aims to define standard, language-
independent RESTful APIs. It enables people and computers to discover and
understand services without accessing source code or documentation or
monitoring network traffic.
• DevCloud consists of the following services:
▫ ProjectMan: provides agile project management and collaboration, supports
management of sprints, milestones, and requirements across projects,
tracks bugs, and provides multi-dimensional statistics reports.
▫ CodeHub: a Git-based online code hosting service for software developers.
It is also a code repository for security management, member and
permission management, branch protection and merging, online editing,
and statistics. The service addresses issues such as cross-region
collaboration, multi-branch concurrent development, and code version
management.
▫ CloudPipeline: provides visualized, customizable pipelines to shorten the
delivery period and improve efficiency.
▫ CodeCheck: manages code quality in the cloud. You can easily perform
static checks and security checks on code in multiple programming
languages and obtain comprehensive quality reports. CodeCheck also
allows you to view grouped defects with fix suggestions provided,
effectively controlling quality.
▫ CloudBuild: provides an easy-to-use hybrid language build platform to
implement cloud-based build, and supports continuous and efficient
delivery. With CloudBuild, you can create, configure, and execute build tasks
with a few clicks to obtain, build, and package code automatically and
monitor build status in real time.
▫ CloudDeploy: provides visualized, one-click deployment. It supports
deployment on VMs or containers by using Tomcat, Spring Boot, and other
templates or by flexibly orchestrating atomic actions. It also supports
parallel deployment and seamless integration with CloudPipeline, providing
standard deployment environments and implementing automatic
deployment.
• E2E process: One platform covers common functions in software development.
These functions are embedded and integrated for governance and O&M.
• Repository locking: You can manually lock a repository to disable any changes or
commits, preventing the stable version to be released from being compromised.
• SSH deployment key: Use the SSH key to control read and write permissions of a
repository. Use the deployment key to enable the read-only permission of a
repository.
• Misoperation tracing and recovery: Code and branches that are deleted by
mistake can be accurately rolled back or retrieved. For deleted repositories,
backups are kept in the physical storage for a specific retention period.
• Operation logs: All operations have tokens. Key operations are audited and
recorded.
• Rule setting: CodeHub allows you to configure commit rules, merge requests, and
gates to ensure that the code quality is controllable.
• High-quality code check rule set based on Huawei's 30-year R&D experience
▫ Provides Java and C/C++ programming guidelines for defect fixing. Provides
automatic fixing of Go code.
• Recommended: ProjectMan, CodeHub, CodeCheck, CloudBuild, CloudDeploy,
CloudTest, CloudArtifact
• ServiceStage provides application hosting, monitoring, alarms, and log analysis
for enterprise developers, test personnel, O&M personnel, and project managers.
The platform is compatible with mainstream application technology stacks,
including multiple languages, microservice frameworks, and running
environments in the industry. It helps enterprises improve the management and
O&M efficiency of traditional, web, and microservice applications, focus on
industry-oriented application innovation, and improve enterprise competitiveness.
• Integrates the software center and archives the built software packages (or
image packages) to the corresponding repositories and organizations.
• Integrates related infrastructure, such as VPC, CCE, ECS, EIP, and ELB. When
deploying applications, you can directly use existing or new infrastructures.
• Integrates the Cloud Service Engine (CSE). You can perform operations related to
microservice governance on the ServiceStage console.
• Integrates storage, database, and cache services and implements persistent data
storage through simple configuration.
• Ever growing services may encounter various unexpected situations, such as
instantaneous and large-scale concurrent access, service errors, and intrusion. The
microservice architecture implements fine-grained service management and
control to meet service requirements.
▫ Supports multiple languages, such as Java, Go, .Node.js, PHP, and Python.
• Answer 2: ABCD
• Discussion 1: Discuss the architecture, development, release, and O&M.
• Console is a visualized entry for cloud resource users to manage and provision
resources.
• Users can the cloud O&M service console and tools to support service O&M.
• With the popularization of microservices, the relationship between applications is
increasingly complex. O&M personnel cannot handle it anymore. Professional
tools are required to comprehensively monitor application calls, and display
service execution traces and statuses, thereby helping users quickly demarcate
performance bottlenecks and faults.
• After applications are migrated to the cloud, users still want microservice
dependency visualization, better end user experience, fast problem tracing,
association analysis on scattered logs. To meet these requirements, Huawei Cloud
provides diverse O&M services to improve O&M efficiency.
• Huawei Cloud launched a dimensional cloud application O&M solution that
integrates AOM and APM. This solution monitors infrastructure, applications, and
services in real time, and supports association analysis of application and
resource alarms, log analysis, intelligent threshold, distributed tracing, and
mobile app exception analysis, enabling users to quickly diagnose and rectify
faults within minutes, and ensure stable application running.
▫ Log management: LTS provides log collection, real-time query, and storage,
helping users easily cope with routine O&M.
▫ Server monitoring provides more than 40 metrics, such as metrics for CPU,
memory, disk, and network, to meet the basic monitoring and O&M
requirements for servers.
▫ CPU usage, memory usage, and the number of opened files used by active
processes give users a better understanding of the ECS or BMS usages.
• Service invoking based on alarm rules are supported. For example, when a
certain type of alarm is triggered, other cloud services (such as FunctionGraph)
can be triggered to perform configured operations.
• Dashboards allow users to compare performance data of different services from
different dimensions. Users must create a dashboard before adding graphs.
• E-commerce services feature large data volume and large data access, which
requires large memory, fast data exchange and processing, and extremely strict
monitoring.
• ECS is a core service in e-commerce scenarios. Therefore, a comprehensive and
three-dimensional ECS monitoring system plays an important role in service
stability. Proactive fine-grained server monitoring of Cloud Eye helps ensure that
e-commerce services run smoothly.
• People access the websites of e-commerce platforms and make transactions.
During grand annual shopping festivals, the websites are often hit by various
problems like slow page loading and long network latency when people access
from different networks. Website monitoring can perform continuous dialing
tests on websites or ECS elastic IP addresses (EIPs) to monitor the availability and
response time of the websites.
• For services used by an e-commerce platform, such as Relational Database
Service (RDS), Elastic Load Balance (ELB), and Virtual Private Cloud (VPC), cloud
service monitoring allows users to track the status of each cloud service and
usage of each metric. After setting alarm rules for cloud service metrics, users can
get a more accurate picture of the health of cloud services.
• An e-commerce platform involves many Huawei Cloud services, such as ECS,
Content Delivery Network (CDN), AS, Web Application Firewall (WAF), RDS, ELB,
and Object Storage Service (OBS). With resource groups, users can view resource
usages, alarms, and health status and manage alarm rules, relating to a specific
service. This greatly reduces O&M complexity and improves O&M efficiency.
• Log auditing is the core of information security audit. They are essential for the
security risk control of information systems in both private and public sectors.
• CTS directly connects to other Huawei Cloud services, records operations on cloud
resources and the results, and transfers these records in the form of trace files to
OBS buckets in real time.
▫ Trace query on the CTS console from the last seven days by multiple
dimensions: trace type, trace source, resource type, filter, operator, trace
status.
▫ Trace file encryption using keys provided by the Data Encryption Workshop
(DEW) during the transfer.
• A trace file is a collection of traces. CTS generates trace files based on services
and transfer cycle and send these files to the specified OBS bucket in real time. In
most cases, all traces of a service generated in a transfer cycle are compressed
into one trace file. However, if there are a large number of traces, CTS will adjust
the number of traces contained in each trace file. Trace files are in JSON format.
• Management trackers record operations on all cloud resources, such as creation,
login, and deletion.
• Log query and real-time analysis: Collected logs can be quickly queried by
keyword or fuzzy match. Users can analyze logs in real time to perform security
diagnosis and analysis, or obtain operations statistics, such as cloud service visits
and clicks.
• Log monitoring and alarm reporting: LTS works with Application Operations
Management (AOM) to count the frequency of specified keywords in logs
retained in LTS. For example, if the keyword ERROR occurs frequently, it can
indicate that services are not running normally.
• Log transfer: Logs of hosts and cloud services are retained in LTS for seven days
by default. Users can also set the retention duration to a value ranging from 1 to
30 days. Retained logs are deleted once the duration is over. For long-term
storage, users can transfer logs to OBS and Data Ingestion Service (DIS).
• Users can configure logs of different types, such as operation logs and access
logs, to be written into different log streams. ICAgent will package and send the
collected logs to LTS by log stream. In this way, users can quickly find the target
logs in the corresponding log streams. The use of log streams greatly reduces the
number of log reads and writes and improves efficiency.
• If ICAgent has been installed on the host for other cloud services, skip the
installation. The time and time zone of the local browser must be consistent with
those of the host before the installation. Users can install ICAgent on the Host
Management page of the LTS console. When ICAgent is installed, users need to
configure log collection paths, which are paths of the host logs to be collected.
• During log structuring, logs with fixed or similar formats are extracted from a log
stream based on the defined structuring method and irrelevant logs are filtered
out. Users can then use SQL syntax to query and analyze the structured logs.
• Collected logs can be quickly queried by keyword or fuzzy match. Users can
analyze logs in real time to perform security diagnosis and analysis, or obtain
operations statistics, such as cloud service visits and clicks.
• Log transfer:
▫ Logs can only be transferred to OBS buckets that are deployed in the same
region as LTS.
• Advantages of AOM:
▫ Open ecosystem: AOM opens O&M data query APIs and collection
standards, and supports independent development.
• Data collection and access layer:
▫ Data transmission: AOM Access is a proxy for receiving O&M data. Received
data will be placed in a Kafka queue. Kafka then transmits the data to the
service computing layer in real time using its high-throughput capability.
▫ Data storage: After being processed by the AOM backend, O&M data is
written into a database. Cassandra stores time series data, Redis is used for
cache query, etcd stores AOM configuration data, and Elasticsearch stores
resources, logs, alarms, and events.
▫ AOM provides basic O&M services such as alarm reporting, logging, and
metric monitoring, and AI services such as exception detection and analysis.
• As cloud migration becomes popular, enterprises are facing the challenge of
managing diverse resources from different cloud vendors. Configuration
management database (CMDB) is a DevOps-based resource management
platform for the entire application lifecycle. As a fundamental service for
automated O&M, it centrally manages the relationships between applications
and resource objects of Huawei Cloud as well as other cloud vendors.
• CMDB functions:
▫ Resource search: Users can search for resources (such as applications and
hosts) by ID, keyword, or name.
• With a dashboard, different graphs can be displayed on the same screen. Various
graphs, such as line graphs, digital graphs, and top N resource graphs allow users
to comprehensively monitor resource data.
• Log search enables users to quickly search for required logs from massive
quantities of logs. Log dump enables users to store logs for a long period of time.
After users create statistical rules, AOM can periodically count keywords in logs
and generate metric data, so that users can monitor system performance and
services in real time. By configuring delimiters, users can divide log content into
multiple words and use these words to search for logs.
• In the cloud era, more and more applications are deployed in the distributed
microservice architecture. As the number of users increases rapidly, many
application exceptions occur. In traditional O&M, metrics cannot be associated
for analysis, so they need manual and subjective processing. This results in low
efficiency, high maintenance costs, and non-ideal performance.
• When there are massive quantities of services, O&M personnel face two major
challenges:
▫ Users choose to leave due to poor experience. O&M personnel fail to detect
and track services with poor experience in real time, and cannot quickly
diagnose application exceptions, greatly affecting user experience.
• 2. C
• Digital twins: Fully utilize the simulation process and completes mapping in the
virtual space to reflect the entire lifecycle of the corresponding physical
equipment, effectively reducing the actual production cost.
• Huawei Cloud EI consists of big data and AI solutions.
• The content in red will be further learned.
• Training framework:
• Computing power:
• MRS provides different big data analysis and processing components for different
scenarios. You can select stream computing components such as Flink for real-
time processing, and Spark or MapReduce for offline batch computing.
▫ Visualized data import and export tool: Use Loader to export data to Data
Warehouse Service (DWS) for business intelligence (BI) analysis.
▫ Storage of mass data: With HBase, you can store a large volume of data
and query data in milliseconds.
▫ Distributed data query: With Spark, you can analyze and query a large
volume of data.
▫ Real-time data ingestion: With Flume, you can achieve real-time data
ingestion and enjoy various data collection and storage access methods.
▫ Data source access: Use Kafka to access the data of tens of thousands of
elevators and escalators in real time.
• Weather data can be stored in OBS and periodically dumped to HDFS for batch
analysis.
• DLI frees you from managing any servers. DLI supports standard SQL and is
compatible with standard SQL and Spark and Flink SQL. It also supports multiple
access modes and mainstream data formats. You can use SQL applications to
query mainstream data formats without data ETL. DLI supports SQL statements
for heterogeneous data sources, including CloudTable, RDS, DWS, CSS, OBS,
custom databases on ECSs, and offline databases.
• DWS is often used together with Cloud Data Migration (CDM) and Data
Ingestion Service (DIS). CDM is used for batch data migration, and DIS is used for
stream data ingestion.
• DataArts Migration: Based on the big data cloud migration and intelligent data
lake solution, DataArts Migration provides easy-to-use migration capabilities and
can integrate a broad set of data sources into the data lake more easily and
efficiently.
• DataArts Architecture can be used to create entity-relationship (ER) models and
dimensional models to standardize and visualize data development and output
data governance methods that can guide development personnel to work with
ease.
• DataArts Factory is a one-stop collaborative big data development platform that
provides fully managed big data scheduling capabilities.
• DataArts Quality can monitor metrics and data quality, and screen out
unqualified data in a timely manner.
• DataArts Catalog provides enterprise-class metadata management to clarify
information assets. It uses a data map to display a data lineage and panorama of
data assets for intelligent data search, operations, and monitoring.
• DataArts DataService enables you to manage APIs centrally and control the
access to subjects, profiles, and metrics. It improves data access, query, and
retrieval efficiency and data consumption experience, and monetizes data assets.
It also allows you to quickly generate new APIs based on data tables, register
your legacy APIs, and centrally manage and publish them.
• DataArts Security provides all-round security assurance to safeguard network
security and control user permissions. It provides a review mechanism for key
processes in DataArts Architecture and DataArts DataService. Data is managed
by level and category throughout the lifecycle, ensuring data privacy compliance
and traceability.
• The long tail is a business strategy that allows companies to realize significant
profits by selling low volumes of hard-to-find items to many customers, instead
of only selling large volumes of a reduced number of popular items.
• Huawei Cloud big data services provide one-stop management and development
throughout the entire data lifecycle and significantly simplify the data
governance process for medium- and long-tail enterprises. With these services,
medium- and long-tail enterprises can analyze a large amount of data more
quickly and efficiently, use data more easily, monetize data in a shorter time, and
digitize their business smoothly.
• As big data has grown, there has been a corresponding growth in the power of
AI. AI has been constantly changing methods of production and how we live.
• AI engineers face many challenges when they are installing and configuring
various AI tools, preparing data, and training models. ModelArts, a one-stop AI
development platform is designed to address these challenges. ModelArts
integrates data preparation, algorithm development, model training, and model
deployment into the production environment, allowing AI engineers to perform
one-stop AI development.
• Data processing: All data formats are supported, as well as team labeling.
• Pangu models: There are multiple foundation models, including the NLP, CV,
multi-modal, and scientific computing models. Through model generalization, the
Pangu models enable large-scale industrialized AI that could not be supported in
traditional AI development. This enables brand-new industrial AI development.