[go: up one dir, main page]

100% found this document useful (1 vote)
8K views14 pages

1Z0-1110-2024 Dumps (Updated Version)

The document outlines various tasks and questions related to Oracle Cloud Infrastructure Data Science, including data labeling formats, custom model training capabilities, and steps for accessing AI services. It also covers topics such as model artifact management, anomaly detection settings, and the use of conda environments. Additionally, it discusses best practices for running machine learning workflows and the Oracle AutoML pipeline sequence.

Uploaded by

Venkataraman
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
8K views14 pages

1Z0-1110-2024 Dumps (Updated Version)

The document outlines various tasks and questions related to Oracle Cloud Infrastructure Data Science, including data labeling formats, custom model training capabilities, and steps for accessing AI services. It also covers topics such as model artifact management, anomaly detection settings, and the use of conda environments. Additionally, it discusses best practices for running machine learning workflows and the Oracle AutoML pipeline sequence.

Uploaded by

Venkataraman
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 14

Oracle Cloud Infrastructure Data Science 2024

Professional (1Z0-1110-24)
1. You have just completed analyzing a set of images by using Oracle Cloud Infrastructure (OCI) Data

Labelling, and you want to export the annotated data. Which two formats are supported?

• CONLL V2003
• COCO
• Data Labelling Service Proprietary JSON
• Spacy

2. As a data scientist, you use the Oracle Cloud Infrastructure (OCI) Language service to train custom

models. Which types of custom models can be trained?

• Image classification, Named Entity Recognition (NER)


• Text classification, Named Entity Recognition (NER)
• Sentiment Analysis, Named Entity Recognition (NER)
• Object detection, Text classification

3. What preparation steps are required to access an Oracle AI service SDK from a Data Science
notebook session?

• Create and upload score.py and runtime.yaml.


• Create and upload the API signing key and config file.
• Import the REST API.
• Call the ADS command to enable AI integration.

4. You are asked to prepare data for a custom-built model that requires transcribing Spanish video

recordings into a readable text format with profane words identified.

Which Oracle Cloud service would you use?

• OCI Translation
• OCI Language
• OCI Speech
• OCI Anomaly Detection

5. You are using Oracle Cloud Infrastructure (OCI) Anomaly Detection to train a model to detect
anomalies in pump sensor data. How does the required False Alarm Probability setting affect an
anomaly detection model?

• It is used to disable the reporting of false alarms.


• It changes the sensitivity of the model to detecting anomalies.
• It determines how many false alarms occur before an error message is generated.
• It adds a score to each signal indicating the probability that its a false alarm.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


6. You are a data scientist leveraging the Oracle Cloud Infrastructure (OCI) Language AI service for
various types of text analyses. Which TWO capabilities can you utilize with this tool?

• Topic classification
• Table extraction
• Sentiment analysis
• Sentence diagramming
• Punctuation correction

7. You want to write a program that performs document analysis tasks such as extracting text and
tables from a document. Which Oracle AI service would you use?

• OCI Language
• Oracle Digital Assistant
• OCI Speech
• OCI Vision

8. Which Oracle Accelerated Data Science (ADS) classes can be used for easy access to data sets from
reference libraries and index websites such as scikit-learn?

• DataLabeling
• DatasetBrowser
• SecretKeeper
• ADSTuner

9. You are a data scientist trying to load data into your notebook session. You understand that
Accelerated Data Science (ADS) SDK supports loading various data formats.

Which of the following THREE are ADS supported data formats?

• DOCX
• Pandas DataFrame
• JSON
• Raw Images
• XML

10. You are a data scientist leveraging Oracle Cloud Infrastructure (OCI) Data Science to create a
model and need some additional Python libraries for processing genome sequencing data. Which of
the following THREE statements are correct with respect to installing additional Python libraries to
process the data?

• You can only install libraries using yum and pip as a normal user.
• You can install private or custom libraries from your own internal repositories.
• OCI Data Science allows root privileges in notebook sessions.
• You can install any open source package available on a publicly accessible Python Package
Index (PyPI) repository.
• You cannot install a library that's not preinstalled in the provided image

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


11. You are a data scientist working for a manufacturing company. You have developed a forecasting
model to predict the sales demand in the upcoming months. You created a model artifact that
contained custom logic requiring third party libraries. When you deployed the model, it failed to run
because you did not include all the third party dependencies in the model artifact. What file should
be modified to include the missing libraries?

• model_artifact_validate.py
• score.py
• requirements.txt
• runtime.yaml

12. You are a data scientist working for a utilities company. You have developed an algorithm that
detects anomalies from a utility reader in the grid. The size of the model artifact is about 2 GB, and
you are trying to store it in the model catalog. Which three interfaces could you use to save the
model artifact into the model catalog?

• Git CLI
• Oracle Cloud Infrastructure (OCI) Command Line Interface (CLI)
• Accelerated Data Science (ADS) Software Development Kit (SDK)
• ODSC CLI
• Console
• OCI Python SDK

13. As a data scientist, you are tasked with creating a model training job that is expected to take
different hyperparameter values on every run. What is the most efficient way to set those
parameters with Oracle Data Science Jobs?

• Create a new job every time you need to run your code and pass the parameters as
environment variables.
• Create a new job by setting the required parameters in your code and create a new job for
every code change.
• Create your code to expect different parameters either as environment variables or as
command line arguments, which are set on every job run with different values.
• Create your code to expect different parameters as command line arguments and create a
new job every time you run the code.

14. You have an embarrassingly parallel or distributed batch job on a large amount of data that you
consider running using Data Science Jobs. What would be the best approach to run the workload?

• Create the job in Data Science Jobs and start a job run. When it is done, start a new job run
until you achieve the number of runs required.
• Create the job in Data Science Jobs and then start the number of simultaneous jobs runs
required for your workload.
• Reconfigure the job run because Data Science Jobs does not support embarrassingly parallel
workloads.
• Create a new job for every job run that you have to run in parallel, because the Data Science
Jobs service can have only one job run per job.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


15. You have received machine learning model training code, without clear information about the
optimal shape to run the training. How would you proceed to identify the optimal compute shape
for your model training that provides a balanced cost and processing time?

• Start with a random compute shape and monitor the utilization metrics and time required to
finish the model training. Perform model training optimizations and performance tests in
advance to identify the right compute shape before running the model training as a job.
• Start with a smaller shape and monitor the Job Run metrics and time required to complete
the model training. If the compute shape is not fully utilized, tune the model parameters,
and re- run the job. Repeat the process until the shape resources are fully utilized.
• Start with the strongest compute shape Job's support and monitor the Job Run metrics and
time required to complete the model training. Tune the model so that it utilizes as much
compute resources as possible, even at an increased cost.
• Start with a smaller shape and monitor the utilization metrics and time required to complete
the model training. If the compute shape is fully utilized, change to compute that has more
resources and re-run the job. Repeat the process until the processing time does not
improve.

16. You have a complex Python code project that could benefit from using Data Science Jobs as it is a
repeatable machine learning model training task. The project contains many subfolders and classes.
What is the best way to run this project as a Job?

• ZIP the entire code project folder and upload it as a Job artifact. Jobs automatically identifies
the_main_ top level where the code is run.
• Rewrite your code so that it is a single executable Python or Bash/Shell script file.
• ZIP the entire code project folder and upload it as a Job artifact on job creation. Jobs
identifies the main executable file automatically.
• ZIP the entire code project folder, upload it as a Job artifact on job creation, and set
JOB_RUN_ENTRYPOINT to point to the main executable file.

17. You are attempting to save a model from a notebook session to the model catalog by using the
Accelerated Data Science (ADS) SDK, with resource principal as the authentication signer, and you
get a 404 authentication error. Which two should you look for to ensure permissions are set up
correctly?

• The model artifact is saved to the block volume of the notebook session.
• A dynamic group has rules that matching the notebook sessions in it compartment.
• The policy for your user group grants manages permissions for the model catalog in this
compartment.
• The policy for a dynamic group grant manages permissions for the model catalog in it
compartment.
• The networking configuration allows access to Oracle Cloud Infrastructure services through a
Service Gateway.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


18. You are a data scientist working inside a notebook session and you attempt to pip install a
package from a public repository that is not included in your conda environment. After running this
command, you get a network timeout error.

What might be missing from your networking configuration?

• FastConnect to an on-premises network.


• Primary Virtual Network Interface Card (VNIC).
• NAT Gateway with public internet access.
• Service Gateway with private subnet access.

19. You are building a model and need input that represents data as morning, afternoon, or evening.
However, the data contains a time stamp. What part of the Data Science life cycle would you be in
when creating the new variable?

• Data access
• Feature engineering
• Model type selection
• Model validation

20. Six months ago, you created and deployed a model that predicts customer churn for a call
centre. Initially, it was yielding quality predictions. However, over the last two months, users are
questioning the credibility of the predictions.

Which two methods would you employ to verify the accuracy of the model?

• Retrain the model


• Validate the model using recent data
• Drift monitoring
• Redeploy the model
• Operational monitoring

21. Which two statements are true about published conda environments?

• They are curated by Oracle Cloud Infrastructure (OCI) Data Science.


• The odac conda init command is used to configure the location of published conda
environments.
• Your notebook session acts as the source to share published conda environments with team
members.
• You can only create a published conda environment by modifying a Data Science conda
environment.
• In addition to service job run environment variables, conda environment variables can be
used in Data Science Jobs.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


22. You have created a conda environment in your notebook session. This is the first time you are
working with published conda environments. You have also created an Object Storage bucket with
permission to manage the bucket.

Which two commands are required to publish the conda environment?

• odac conda publish --slug <SLUG>


• odsc conda list --override
• odsc conda init --bucket_namespace <NAMESPACE> --bucket_name <BUCKET>
• odsc conda create --file manifest.yaml
• conda activate /home/datascience/conda/<SLUG>/

23. When preparing your model artifact to save it to the Oracle Cloud Infrastructure (OCI) Data
Science model catalog, you create a score.py file. What is the purpose of the score.py file?

• Configure the deployment infrastructure.


• Execute the inference logic code.
• Define the compute scaling strategy.
• Define the inference server dependencies.

24. You realize that your model deployment is about to reach its utilization limit. What would you do
to avoid the issue before requests start to fail?

• Update the deployment to add more instances.


• Delete the deployment.
• Update the deployment to use fewer instances.
• Update the deployment to use a larger virtual machine (more CPUs/memory).
• Reduce the load balancer bandwidth limit so that fewer requests come in.

25. You are working as a data scientist for a healthcare company. They decide to analyze the data to
find patterns in a large volume of electronic medical records. You are asked to build a PySpark
solution to analyze these records in a JupyterLab notebook. What is the order of recommended
steps to develop a PySpark application in Oracle Cloud Infrastructure (OCI) Data Science?

• Launch a notebook session. Install a PySpark conda environment. Configure core-site.xml.


Develop your PySpark application. Create a Data Flow application with the Accelerated Data
Science (ADS) SDK.
• Install a Spark conda environment. Configure core-site.xml. Launch a notebook session.
Create a Data Flow application with the Accelerated Data Science (ADS) SDK. Develop your
PySpark application.
• Configure core-site.xml. Install a PySpark conda environment. Create a Data Flow application
with the Accelerated Data Science (ADS) SDK. Develop your PySpark application. Launch a
notebook session.
• Launch a notebook session. Configure core-site.xml. Install a PySpark conda environment.
Develop your PySpark application. Create a Data Flow application with the Accelerated Data
Science (ADS) SDK.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


26. Which Oracle Cloud Infrastructure (OCI) service should you use to create and run Spark
applications using ADS?

• Data Integration
• Vault
• Data Flow
• Analytics Cloud

27. For your next data science project, you need access to public geospatial images.

Which Oracle Cloud service provides free access to those images?

• Oracle Open Data


• Oracle Big Data Service
• Oracle Cloud Infrastructure Data Science
• Oracle Analytics Cloud

28. You are a data scientist building a pipeline in the Oracle Cloud Infrastructure (OCI) Data Science
service for your machine learning project. You want to optimize the pipeline completion time by
running some steps in parallel. Which statement is true about running pipeline steps in parallel?

• Steps in a pipeline can be run only sequentially.


• Pipeline steps can be run in sequence or in parallel, as long as they create a directed acyclic
graph (DAG).
• All pipeline steps are always run in parallel.
• Parallel steps cannot be run if they are completely independent of each other.

29. You want to build a multistep machine learning workflow by using the Oracle Cloud
Infrastructure (OCI) Data Science Pipeline feature. How would you configure the conda environment
to run a pipeline step?

• Configure a compute shape.


• Configure a block volume.
• Use command-line variables.
• Use environmental variables.

30. You want to write a Python script to create a collection of different projects for your data science
team. Which Oracle Cloud Infrastructure (OCI) Data Science interface would you use?

• The OCI Software Development Kit (SDK)


• OCI Console
• Command line interface (CLI)
• Mobile App

31. You are a data scientist designing an air traffic control model, and you choose to leverage Oracle
AutoML You understand that the Oracle AutoML pipeline consists of multiple stages and
automatically operates in a certain sequence. What is the correct sequence for the Oracle AutoML
pipeline?

• Algorithm selection, Feature selection, Adaptive sampling, Hyperparameter tuning

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


• Adaptive sampling, Algorithm selection, Feature selection, Hyperparameter tuning
• Adaptive sampling, Feature selection, Algorithm selection, Hyperparameter tuning
• Algorithm selection, Adaptive sampling, Feature selection, Hyperparameter tuning

32. You have trained three different models on your data set using Oracle AutoML. You want to
visualize the behavior of each of the models, including the baseline model, on the test set. Which
class should be used from the Accelerated Data Science (ADS) SDK to visually compare the models?

• EvaluationMetrics
• ADSEvaluator
• ADSExplainer
• ADSTuner

33. Using Oracle AutoML, you are tuning hyperparameters on a supported model class and have
specified a time budget. AutoML terminates computation once the time budget is exhausted. What
would you expect AutoML to return in case the time budget is exhausted before hyperparameter
tuning is completed?

• The current best-known hyperparameter configuration is returned.


• A random hyperparameter configuration is returned.
• A hyperparameter configuration with a minimum learning rate is returned.
• The last generated hyperparameter configuration is returned.

34. As a data scientist, you are trying to automate a machine learning (ML) workflow and have
decided to use Oracle Cloud Infrastructure (OCI) AutoML Pipeline.

Which three are part of the AutoML Pipeline?

• Feature Selection
• Adaptive Sampling
• Model Deployment
• Feature Extraction
• Algorithm Selection

35. You want to use ADSTuner to tune the hyperparameters of a supported model you recently
trained. You have just started your search and want to reduce the computational cost as well as
access the quality of the model class that you are using.

What is the most appropriate search space strategy to choose?

• Detailed
• ADSTuner doesn't need a search space to tune the hyperparameters.
• Perfunctory
• Pass a dictionary that defines a search space.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


36. You have just received a new data set from a colleague. You want to quickly find out summary
information about the data set, such as the types of features, the total number of observations, and
distributions of the data. Which Accelerated Data Science (ADS) SDK method from the ADSDataset
class would you use?

• show_corr()
• to_xgb ()
• compute ()
• show_in_notebook ()

37. You want to make your model more frugal to reduce the cost of collecting and processing data.
You plan to do this by removing features that are highly correlated. You would like to create a heat
map that displays the correlation so that you can identify candidate features to remove.

Which Accelerated Data Science (ADS) SDK method is appropriate to display the comparability
between Continuous and Categorical features?

• pearson_plot()
• cramersv_plot()
• correlation_ratio_plot()
• corr()

38. You have built a machine model to predict whether a bank customer is going to default on a
loan. You want to use Local Interpretable Model-Agnostic Explanations (LIME) to understand a
specific prediction. What is the key idea behind LIME?

• Global behaviour of a machine learning model may be complex, while the local behaviour
may be approximated with a simpler surrogate model.
• Model-agnostic techniques are more interpretable than techniques that are dependent on
the types of models.
• Global and local behaviours of machine learning models are similar.
• Local explanation techniques are model-agnostic, while global explanation techniques are
not.

39. You want to evaluate the relationship between feature values and target variables. You have a
large number of observations having a near uniform distribution and the features are highly
correlated.

Which model explanation technique should you choose?


• Feature Permutation Importance Explanations
• Local Interpretable Model-Agnostic Explanations
• Feature Dependence Explanations
• Accumulated Local Effects

40. As you are working in your notebook session, you find that your notebook session does not have
enough compute CPU and memory for your workload.

How would you scale up your notebook session without losing your work?

• Create a temporary bucket on Object Storage, write all your files and data to Object Storage,
delete your notebook session, provision a new notebook session on a larger compute shape,
Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)
and copy your files and data from your temporary bucket onto your new notebook session.
• Ensure your files and environments are written to the block volume storage under the
/home/datascience directory, deactivate the notebook session, and activate the notebook
session with a larger compute shape selected.
• Download all your files and data to your local machine, delete your notebook session,
provision a new notebook session on a larger compute shape, and upload your files from
your local machine to the new notebook session.
• Deactivate your notebook session, provision a new notebook session on a larger compute
shape and re-create all of your file changes.

41. You are creating an Oracle Cloud Infrastructure (OCI) Data Science job that will run on a recurring
basis in a production environment. This job will pick up sensitive data from an Object Storage
bucket, train a model, and save it to the model catalog.

How would you design the authentication mechanism for the job?

• Create a pre-authenticated request (PAR) for the Object Storage bucket, and use that in the
job code.
• Use the resource principal of the job run as the signer in the job code, ensuring there is a
dynamic group for this job run with appropriate access to Object Storage and the model
catalog.
• Store your personal OCI config file and keys in the Vault and access the Vault through the job
run resource principal.
• Package your personal OCI config file and keys in the job artifact.

42. You have created a Data Science project in a compartment called Development and shared it
with a group of collaborators. You now need to move the project to a different compartment called
Production after completing the current development iteration.

Which statement is correct?

• Moving a project to a different compartment also moves its associated notebook sessions
and models to the new compartment.
• Moving a project to a different compartment requires deleting all its associated notebook
sessions and models first.
• You cannot move a project to a different compartment after it has been created.
• You can move a project to a different compartment without affecting its associated
notebook sessions and models.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


43. Select two reasons why it is important to rotate encryption keys when using Oracle Cloud
Infrastructure (OCI) Vault to store credentials or other secrets.

• Key rotation allows you to encrypt no more than five keys at a time.
• Key rotation improves encryption efficiency.
• Periodically rotating keys make it easier to reuse keys.
• Key rotation reduces risk if a key is ever compromised.
• Periodically rotating keys limits the amount of data encrypted by one key version.

44. As a data scientist, you have stored sensitive data in a database. You need to protect this data by
using a master encryption algorithm, which uses symmetric keys. Which master encryption
algorithm would you choose in the Oracle Cloud Infrastructure (OCI) Vault service?

• Triple Data Encryption Standard Algorithm


• Elliptical Curve Cryptography Digital Signature Algorithm
• Advanced Encryption Standard Keys
• Rivert-Shamir-Adleman Keys

45. As a data scientist, you create models for cancer prediction based on mammographic images.
The correct identification is very crucial in this case. After evaluating two models, you arrive at the
following confusion matrix.

• Model 1 has Test accuracy is 80% and recall is 70%.

• Model 2 has Test accuracy is 75% and recall is 85%.

Which model would you prefer and why?

• Model 2, because recall is high.


• Model 1, because the test accuracy is high.
• Model 2, because recall has more impact on predictions in this use se.
• Model 1, because recall has lesser impact on predictions in this use case.

46. You train a model to predict housing prices for your city. Which two metrics from the
Accelerated Data Science (ADS) ADSEvaluator class can you use to evaluate the regression model?

• Explained Variance Score


• F-1 Score
• Weighted Precision
• Weighted Recall
• Mean Absolute Error

47. In the Oracle Cloud Infrastructure (OCI) Data Science service, how does Model Catalog help with
model deployment and management in MLOps?

• It is a database that stores all the features used in a machine learning model.
• It helps to automate the feature engineering process.
• It provides a centralized and scalable way to manage models and their metadata.
• It helps to package the model and its dependencies into a lightweight, portable container.

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


48. You want to ensure that all stdout and stderr from your code are automatically collected and
logged, without implementing additional logging in your code. How would you achieve this with Data
Science Jobs?

• On job creation, enable logging and select a log group. Then, select either a log or the option
to enable automatic log creation.
• Make sure that your code is using the standard logging library and then store all the logs to
Object Storage at the end of the job.
• Create your own log group and use a third-party logging service to capture job run details for
log collection and storing.
• You can implement custom logging in your code by using the Data Science Jobs logging
service.

49. As a data scientist, you are working on a global health data set that has data from more than 50
countries. You want to encode three features such as 'countries', 'race' and 'body organ' as
categories.

Which option would you use to encode the categorical feature?

• OneHotEncoder ()
• DataFrameLabelEncoder ()
• show_in_notebook ()
• auto_transform()

50. While reviewing your data, you discover that your data set has a class imbalance. You are aware
that the Accelerated Data Science (ADS) SDK provides multiple built-in automatic transformation
tools for data set transformation. Which would be the right tool to correct any imbalance between
the classes?

• visualize_transforms ()
• auto_transform()
• sample ()
• suggest_recommendations ()

51. After you have created and opened a notebook session, you want to use the Accelerated Data
Science (ADS) SDK to access your data and get started with an exploratory data analysis.

From which two places can you access or install the ADS SDK?

• Oracle Autonomous Data Warehouse


• Oracle Machine Learning (OML)
• Oracle Big Data Service
• Conda environments in Oracle Cloud Infrastructure (OCI) Data Science
• Python Package Index (PyPI)

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


52. You loaded data into Oracle Cloud Infrastructure (OCI) Data Science. To transform the data, you
want to use the Accelerated Data Science (ADS) SDK. When you applied the get_recommendations ()
tool to the ADSDataset object, it showed you user-detected issues with all the recommended
changes to apply to the dataset. Which option should you use to apply all the recommended
transformations at once?

• get_transformed_dataset ()
• fit_transform()
• auto_transform()
• visualize_transforms ()

53. The Accelerated Data Science (ADS) model evaluation classes support different types of machine
learning modeling techniques. Which three types of modeling techniques are supported by ADS
Evaluators?

• Principal Component Analysis


• Multiclass Classification
• K-means Clustering
• Recurrent Neural Network
• Binary Classification
• Regression Analysis

54. During a job run, you receive an error message that no space is left on your disk device. To solve
the problem, you must increase the size of the job storage. What would be the most efficient way to
do this with Data Science Jobs?

• Create a new job with increased storage size and then run the job.
• On the job run, set the environment variable that helps increase the size-of the storage.
• Your code is using too much disk space. Refactor the code to identify the problem.
• Edit the job, change the size of the storage of your job, and start a new job run.

55. You have trained a machine learning model on Oracle Cloud Infrastructure (OCI) Data Science,
and you want to save the code and associated pickle file in a Git repository. To do this, you have to
create a new SSH key pair to use for authentication. Which SSH command would you use to create
the public/private algorithm key pair in the notebook session?

• ssh-agent
• ssh-copy-id
• ssh-add
• ssh-Keygen

Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)


Want any exam dump in pdf email me at info.tipsandtricks10@gmail.com (Little Paid)

You might also like