0% found this document useful (0 votes)
69 views5 pages

CCD CT 2 Model Answer 2022-23

CCD
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
69 views5 pages

CCD CT 2 Model Answer 2022-23

CCD
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 5

ALL INDIA SHRI SHIVAJI MEMORIAL SOCIETY’S POLYTECHNIC, PUNE – 01

MODEL ANSWER
Class Test- I (2023-24)

Program Code :- AN5I Semester :- V


Course Title :- Cloud Computing with data Science Course code:- 22594
Marks :- 20 Time :- 1 Hrs.
Artificial intelligence and machine
Program :- Date :-
leaning

Course Outcomes:- C22594.4, C22594.5, C22594.6

Sub.
Que. Total
Que Stepwise Solution Marks
No. Marks
.
1 Attempt any FOUR (C22594.1) 2 8
a Define Data Pipeline 02 02
Define: Define
1. A data pipeline is a method in which raw data is ingested from various data sources and 2M
then ported to data store, like a data lake or data warehouse, for analysis.
2. A data pipeline is a series of processes that migrate data from a source to a destination
database. An example of a technical dependency may be that after assimilating data from
sources, the data is held in a central queue before subjecting it to further validations and
then finally dumping into a destination.
b State the use of Modern data Pipeline. 02 02
1.Data Pipeline is a web service that helps you reliably process and move data between
different AWS compute and storage services, as well as on-premises data sources, at For
specified intervals. each
2.data pipelines provide the foundation for a range of data projects; this can include use – 1
exploratory data analyses, data visualizations, and machine learning tasks. Mark
3.automation of data pipelines allows organizations to extract data at its source,
transform it, integrate it with other sources and fuel business applications and data
analytics
c Describe Any Four Characteristics of Data Pipeline . 02 02
Characteristics OF Data pipeline: For
1.Continuous, extensible data processing each
2.Cloud-enabled elasticity and agility Charact
3.Independent, isolated data processing resources eristics
4.Widespread data access and the ability to self-serve – 1/2
5.High availability and disaster recovery Mark

d Enlist the process Of Designing pipeline. 02 02

Process Of Designing the pipeline in Cloud: Proper


process
Step 1: Determine the goal 2M
When designing a data pipeline, the priority is to identify the outcome or value the
data pipeline will bring to your company or product.
Step 2: Choose the data sources
We then consider the possible data sources that’ll enter the data pipeline .
Step 3: Determine the data ingestion strategy
With the pipeline goal and data sources understood. We need to ask questions
about how the pipeline will collect the data.
Step 4: Design the data processing plan
Once data has been ingested, it has to be processed and transformed for it to be
valuable to downstream systems
Step 5: Set up storage for the output of the pipeline
Once the data has been processed, we must determine the final storage
destination for our data to serve various business use cases.
Step 6: Plan the data workflow
We then need to design the sequencing of processes in the data pipeline.
Step 7: Implement a data monitoring and governance framework
In this step, we establish a data monitoring and governance framework, which
helps us observe the data pipeline to ensure a healthy and efficient channel that’s
reliable, secure, and performs as required.
Step 8: Plan the data consumption layer
This final step determines the various services that’ll consume the processed data
from our data pipeline.
e Explain ETL Process in detail(U) 02 02
1. ETL is a process in Data Warehousing and it stands for Extract, Transform and Any 2
Load. It is a process in which an ETL tool extracts the data from various data source for
systems, transforms it in the staging area, and then finally, loads it into the Data Each 1
Warehouse system. M
2.Extract, transform, and load (ETL) is the process of combining data from multiple
sources into a large, central repository called a data warehouse. ETL uses a set of
business rules to clean and organize raw data and prepare it for storage, data
analytics, and machine learning (ML).

2 A Attempt any TWO (C22319.4) 2 8


a List any four Common issues in Kubernetes
Issues in kubernetes: Any 4
1. Complexity: Kubernetes has a steep learning curve. Setting up, configuring, and four
maintaining a Kubernetes cluster can be challenging, especially for those new to container issues
orchestration. for
2. Resource Costs: While Kubernetes can optimize resource usage, running Kubernetes each
clusters can still be expensive in terms of cloud infrastructure costs. 1M
3. Management Overhead: Operating and managing Kubernetes clusters require ongoing
effort and expertise. This includes tasks like updates, monitoring, and troubleshooting.
4. Networking Complexity: Networking in Kubernetes can be complex, and setting up
communication between services or across clusters can be challenging, especially in a
multi-cloud setup.
5. Lack of Application Visibility: Kubernetes abstracts many infrastructure details, which
can make it difficult to get comprehensive insights into the performance and behavior of
applications.
6. Vendor Lock-In: While Kubernetes promotes portability, some cloud providers offer
managed Kubernetes services (e.g., Amazon EKS, Google GKE) with cloud-specific features,
potentially creating vendor lock-in.
7. Data Management: Handling stateful applications and managing persistent data in a
Kubernetes environment can be challenging.
8. Version Compatibility: Ensuring that your application and its dependencies are
compatible with the Kubernetes version you're using can be a concern, especially when
upgrading
b Draw and Explain Docker Architecture in detail. 4 4

Docker architecture:- Diagra


m
Docker uses a client-server architecture. The Docker client talks to the Docker
1M,Ex
daemon, which does the heavy lifting of building, running, and distributing your
plainati
Docker containers. The Docker client and daemon can run on the same system, or on 3M
you can connect a Docker client to a remote Docker daemon. The Docker client
and daemon communicate using a REST API, over UNIX sockets or a network
interface. Another Docker client is Docker Compose, that lets you work with
applications consisting of a set of containers.

1.The Docker daemon


The Docker daemon (dockerd) listens for Docker API requests and manages Docker
objects such as images, containers, networks, and volumes. A daemon can also
communicate with other daemons to manage Docker services.
2.The Docker client
The Docker client (docker) is the primary way that many Docker users interact with
Docker. When you use commands such as docker run, the client sends these
commands to dockerd, which carries them out. The docker command uses the
Docker API. The Docker client can communicate with more than one daemon.
3.Docker Desktop
Docker Desktop is an easy-to-install application for your Mac, Windows or Linux
environment that enables you to build and share containerized applications and
microservices. Docker Desktop includes the Docker daemon (dockerd), the Docker
client (docker), Docker Compose, Docker Content Trust, Kubernetes, and
Credential Helper. For more information, see Docker Desktop.
4.Docker registries
A Docker registry stores Docker images. Docker Hub is a public registry that
anyone can use, and Docker looks for images on Docker Hub by default. You can
even run your own private registry.When you use the docker pull or docker
run commands, Docker pulls the required images from your configured registry.
When you use the docker push command, Docker pushes your image to your
configured registry.
5.Docker objects
When you use Docker, you are creating and using images, containers, networks,
volumes, plugins, and other objects. This section is a brief overview of some of
those objects.
6.Images
An image is a read-only template with instructions for creating a Docker container.
Often, an image is based on another image, with some additional customization.
For example, you may build an image which is based on the ubuntu image, but
installs the Apache web server and your application, as well as the configuration
details needed to make your application run.
7.Containers
A container is a runnable instance of an image. You can create, start, stop, move,
or delete a container using the Docker API or CLI.
c Explain Elastic Resource in detail 4
1. Elastic resources in cloud computing refer to the ability to dynamically and Explain 4
automatically scale computing resources up or down based on the changing demands of ation
an application or workload. 4M
2. This elasticity allows cloud users to efficiently allocate and de-allocate resources as
needed, which can help optimize performance, cost, and resource utilization.
3. Elasticity is a fundamental feature of cloud computing and is particularly valuable in
scenarios where workloads are unpredictable or vary over time.
4. It enables organizations to optimize their infrastructure in a cost-effective and
responsive manner, supporting the efficient use of cloud resources.
B Attempt any ONE (C22594.3) 4 4
a Explain AWS SageMaker with example.(. 4
AWS Sagemaker:
Explain
ation
4M
1. Amazon SageMaker is a fully managed machine learning service. With SageMaker, data
scientists and developers can quickly and easily build and train machine learning models,
and then directly deploy them into a production-ready hosted environment.
2. It provides an integrated Jupyter authoring notebook instance for easy access to your
data sources for exploration and analysis, so you don't have to manage servers.
3. It also provides common machine learning algorithms that are optimized to run
efficiently against extremely large data in a distributed environment.
4. With native support for bring-your-own-algorithms and frameworks, SageMaker offers
flexible distributed training options that adjust to your specific workflows.
5.Deploy a model into a secure and scalable environment by launching it with a few clicks
from SageMaker Studio or the SageMaker console.
Features Of Sagemaker:
SageMaker geospatial capabilities
Build, train, and deploy ML models using geospatial data.
SageMaker Model Cards
Document information about your ML models in a single place for streamlined
governance and reporting throughout the ML lifecycle.
SageMaker Model Dashboard
A pre-built, visual overview of all the models in your account. Model Dashboard
integrates information from SageMaker Model Monitor, transform jobs,
endpoints, lineage tracking, and CloudWatch so you can access high-level model
information and track model performance in one unified view.
SageMaker Role Manager
Administrators can define least-privilege permissions for common ML activities
using custom and preconfigured persona-based IAM roles.
AutoML step
Create an AutoML job to automatically train a model in SageMaker Pipelines.
Collaboration with shared spaces
A shared space consists of a shared JupyterServer application and a shared
directory. All user profiles in a Domain have access to all shared spaces in the
Domain.
Data Wrangler data preparation widget
Interact with your data, get visualizations, explore actionable insights, and fix data
quality issues.
Inference shadow tests
Evaluate any changes to your model-serving infrastructure by comparing its
performance against the currently deployed infrastructure.
b State any Four ML systems available in Market . 4
ML systems available in Market For
each
Chatbots: Chatbots are available on an organization's website and serve as 1M
customer support. Intelligent bots are able to deduce customer queries by asking a
strategic set of questions. These are available as plug-and-play packages that fit
right into the web page via an endpoint.

Recommender Systems: Recommendation systems use clever statistics to profile


customers and provide them with relevant recommendations. These may take
some time to adjust to your customers' data but can be easily built.

Sentiment Analysis: Marketers use these systems to predict customer sentiments


regarding advertisements and products. These analyses can be used to improve
experience and customer retention.

Fraud detection: Another prominent use of machine learning in business is in


fraud detection, particularly in banking and financial services, where institutions
use it to alert customers of potentially fraudulent use of their credit and debit
cards.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy