Weekend Sale Limited Time Flat 70% Discount offer - Ends in 0d 00h 00m 00s - Coupon code: 70spcl

Oracle 1z0-1110-25 Oracle Cloud Infrastructure 2025 Data Science Professional Exam Practice Test

Page: 1 / 16
Total 158 questions

Oracle Cloud Infrastructure 2025 Data Science Professional Questions and Answers

Question 1

In which two ways can you improve data durability in Oracle Cloud Infrastructure Object Storage?

Options:

A.

Setup volumes in a RAID1 configuration

B.

Enable server-side encryption

C.

Enable Versioning

D.

Limit delete permissions

E.

Enable client-side encryption

Question 2

You are a data scientist; you use the Oracle Cloud Infrastructure (OCI) Language service to train custom models. Which types of custom models can be trained?

Options:

A.

Image classification, Named Entity Recognition (NER)

B.

Text classification, Named Entity Recognition (NER)

C.

Sentiment Analysis, Named Entity Recognition (NER)

D.

Object detection, Text classification

Question 3

Which is NOT a valid OCI Data Science notebook session approach?

Options:

A.

Ensure you don’t execute long-running Python processes in a notebook cell

B.

Run the process directly in the terminal and use Python logging to get updates on the progress of your job

C.

Avoid having multiple users in the same notebook session due to the possibility of resource contention and write conflicts

D.

While connecting to data in OCI Object Storage from your notebook session, the best practice is to make a local copy on the device and then upload it to your notebook session block volume

E.

Authenticate using your notebook session’s resource principal to access other OCI resources. Resource principals provide a more secure way to authenticate to resources compared to the OCI configuration and API approach

Question 4

You have an embarrassingly parallel or distributed batch job with a large amount of data running using Data Science Jobs. What would be the best approach to run the workload?

Options:

A.

Create a job in Data Science Jobs and then start the number of simultaneous job runs required for your workload

B.

Create a new job for every job run that you have to run in parallel, because the Data Science Job service can have only one job per job

C.

Create the job in Data Science Jobs and start a job run. When it is done, start a new job run until you achieve the number of runs required

D.

Reconfigure the job run because Data Science Jobs does not support embarrassingly parallel

Question 5

Which architecture is based on the principle of “never trust, always verify”?

Options:

A.

Federated identity

B.

Zero trust

C.

Fluid perimeter

D.

Defense in depth

Question 6

During a job run, you receive an error message that no space is left on your disk device. To solve the problem, you must increase the size of the job storage. What would be the most efficient way to do this with Data Science Jobs?

Options:

A.

Edit the job, change the size of the storage of your job, and start a new job run

B.

On the job run, set the environment variable that helps increase the size of the storage

C.

Create a new job with increased storage size and then run the job

D.

Your code using too much disk space. Refactor the code to identify the problem

Question 7

As a data scientist, you are tasked with creating a model training job that is expected to take different hyperparameter values on every run. What is the most efficient way to set those parameters with Oracle Data Science Jobs?

Options:

A.

Create a new job every time you need to run your code and pass the parameters as environment variables

B.

Create your code to expect different parameters as command-line arguments, and create a new job every time you run the code

C.

Create a new job by setting the required parameters in your code, and create a new job for every code change

D.

Create your code to expect different parameters either as environment variables or as command-line arguments, which are set on every job run with different values

Question 8

Which statement about Oracle Cloud Infrastructure Data Science Jobs is true?

Options:

A.

Jobs provisions the infrastructure to run a process on-demand

B.

Jobs comes with a set of standard tasks that cannot be customized

C.

You must create and manage your own Jobs infrastructure

D.

You must use a single Shell/Bash or Python artifact to run a job

Question 9

As a data scientist, you are tasked with creating a model training job that is expected to take different hyperparameter values on every run. What is the most efficient way to set those parameters with Oracle Data Science Jobs?

Options:

A.

Create a new job every time you need to run your code and pass the parameters as environment variables

B.

Create a new job by setting the required parameters in your code and create a new job for every code change

C.

Create your code to expect different parameters either as environment variables or as command-line arguments, which are set on every job run with different values

D.

Create your code to expect different parameters as command-line arguments and create a new job every time you run the code

Question 10

Which OCI service provides a managed Kubernetes service for deploying, scaling, and managing containerized applications?

Options:

A.

Oracle Cloud Infrastructure Container Registry

B.

Oracle Cloud Infrastructure Load Balancing

C.

Oracle Cloud Infrastructure Container Engine for Kubernetes

D.

Oracle Cloud Infrastructure Streaming

Question 11

Which feature of Oracle Cloud Infrastructure Data Science provides an interactive coding environment for building and training machine learning models?

Options:

A.

Model Catalog

B.

Jobs

C.

Notebook Sessions

D.

Projects

Question 12

Which CLI command allows the customized conda environment to be shared with co-workers?

Options:

A.

odsc conda clone

B.

odsc conda publish

C.

odsc conda modify

D.

odsc conda install

Question 13

Which OCI service enables you to build, train, and deploy machine learning models in the cloud?

Options:

A.

Oracle Cloud Infrastructure Data Catalog

B.

Oracle Cloud Infrastructure Data Integration

C.

Oracle Cloud Infrastructure Data Science

D.

Oracle Cloud Infrastructure Data Flow

Question 14

You want to write a Python script to create a collection of different projects for your data science team. Which Oracle Cloud Infrastructure (OCI) Data Science interface would you use?

Options:

A.

The OCI Software Development Kit (SDK)

B.

OCI Console

C.

Command Line Interface (CLI)

D.

Mobile App

Question 15

You have received machine learning model training code, without clear information about the optimal shape to run the training. How would you proceed to identify the optimal compute shape for your model training that provides a balanced cost and processing time?

Options:

A.

Start with a smaller shape and monitor the Job Run metrics and time required to complete the model training. If the compute shape is not fully utilized, tune the model parameters, and rerun the job. Repeat the process until the shape resources are fully utilized

B.

Start with the strongest compute shape Jobs support and monitor the Job Run metrics and time required to complete the model training. Tune the model so that it utilizes as much compute resources as possible, even at an increased cost

C.

Start with a smaller shape and monitor the utilization metrics and time required to complete the model training. If the compute shape is fully utilized, change to compute that has more resources and rerun the job. Repeat the process until the processing time does not improve

D.

Start with a random compute shape and monitor the utilization metrics and time required to finish the model training. Perform model training optimizations and performance tests in advance to identify the right compute shape before running the model training as a job

Question 16

True or false? Bias is a common problem in data science applications.

Options:

A.

True

B.

False

Question 17

You have received machine learning model training code, without clear information about the optimal shape to run the training on. How would you proceed to identify the optimal compute shape for your model training that provides a balanced cost and processing time?

Options:

A.

Start with a smaller shape and monitor the job run metrics and time required to complete the model training. If the compute shape is not fully utilized, tune the model parameters, and rerun the job. Repeat the process until the shape resources are fully utilized.

B.

Start with the strongest compute shape Jobs support and monitor the job run metrics and time required to complete the model training. Tune the model so that it utilizes as much compute resources as possible, even at an increased cost.

C.

Start with a small shape and monitor the utilization metrics and time required to complete the model training. If the compute shape is fully utilized, change to compute that has more resources and rerun the job. Repeat the process until the processing time does not improve.

D.

Start with a random compute shape and monitor the utilization metrics and time required to finish the model training. Perform model training optimization and performance tests in advance to identify the right compute shape before running the model training as a job.

Question 18

Which function's objective is to represent the difference between the predictive value and the target value?

Options:

A.

Optimizer function

B.

Fit function

C.

Update function

D.

Cost function

Question 19

Select two reasons why it is important to rotate encryption keys when using Oracle Cloud Infrastructure (OCI) Vault to store credentials or other secrets.

Options:

A.

Key rotation allows you to encrypt no more than five keys at a time

B.

Key rotation improves encryption efficiency

C.

Periodically rotating keys makes it easier to reuse keys

D.

Key rotation reduces risk if a key is ever compromised

E.

Periodically rotating keys limits the amount of data encrypted by one key version

Question 20

What is the name of the machine learning library used in Apache Spark?

Options:

A.

MLib

B.

GraphX

C.

Structured Streaming

D.

HadoopML

Question 21

What is a common maxim about data scientists?

Options:

A.

They spend 80% of their time finding and preparing data and 20% analyzing it.

B.

They spend 80% of their time analyzing data and 20% finding and preparing it.

C.

They spend 80% of their time on failed analytics projects and 20% doing useful work.

Question 22

You are a data scientist trying to load data into your notebook session. You understand that Accelerated Data Science (ADS) SDK supports loading various data formats. Which of the following THREE are ADS-supported data formats?

Options:

A.

DOCX

B.

Pandas DataFrame

C.

JSON

D.

Raw Images

E.

XML

Question 23

What is feature engineering in machine learning used for?

Options:

A.

To perform parameter tuning

B.

To interpret ML models

C.

To transform existing features into new ones

D.

To help understand the dataset features

Question 24

Arrange the following in the correct Git Repository workflow order:

    Install, configure, and authenticate Git.

    Configure SSH keys for the Git repository.

    Create a local and remote Git repository.

    Commit files to the local Git repository.

    Push the commit to the remote Git repository.

Options:

A.

2, 3, 1, 4, 5

B.

4, 2, 3, 1, 5

C.

3, 5, 1, 2, 4

D.

1, 2, 3, 4, 5

Question 25

What is a conda environment?

Options:

A.

A system that manages package dependencies

B.

A collection of kernels

C.

An open-source environment management system

D.

An environment deployment system on Oracle AI

Question 26

You want to evaluate the relationship between feature values and target variables. You have a large number of observations having a near uniform distribution and the features are highly correlated. Which model explanation technique should you choose?

Options:

A.

Feature Permutation Importance Explanations

B.

Local Interpretable Model-Agnostic Explanations

C.

Feature Dependence Explanations

D.

Accumulated Local Effects

Question 27

Which encryption is used for Oracle Data Science?

Options:

A.

256-bit Advanced Encryption Standard (AES-256)

B.

Data Encryption Standard (DES)

C.

Triple DES (TDES)

D.

Twofish

E.

Rivest Shamir Adleman (RSA)

Question 28

Which Security Zone policy is NOT valid?

Options:

A.

A boot volume can be moved from a security zone to a standard compartment

B.

A compute instance cannot be moved from a security zone to a standard compartment

C.

Resources in a security zone should not be accessible from the public internet

D.

Resources in a security zone must be automatically backed up regularly

Question 29

Which of the following analytical and statistical techniques do data scientists commonly use?

Options:

A.

Classification

B.

Regression

C.

Clustering

D.

All of the above

Question 30

Which statement is true about origin management in Web Application Firewall (WAF)?

Options:

A.

Multiple origins can be defined

B.

Only a single origin can be active for a WAF

C.

Only statement B is true

D.

Both the statements are false

E.

Both the statements are true

F.

Only statement A is true

Question 31

Which THREE types of data are used for Data Labeling?

Options:

A.

Audio

B.

Text Document

C.

Images

D.

Graphs

Question 32

Which model has an open-source, open model format that allows you to run machine learning models on different platforms?

Options:

A.

PySpark

B.

PyTorch

C.

TensorFlow

D.

ONNX

Question 33

Which statement about Oracle Cloud Infrastructure Anomaly Detection is true?

Options:

A.

Accepted file types are SQL and Python

B.

Data used for analysis can be text or numerical in nature

C.

It is an important tool for detecting fraud, network intrusions, and discrepancies in sensor time series analysis

D.

It is trained on a combination of customer and general industry datasets

Question 34

For your next data science project, you need access to public geospatial images. Which Oracle Cloud service provides free access to those images?

Options:

A.

Oracle Big Data Service

B.

Oracle Analytics Cloud

C.

Oracle Cloud Infrastructure (OCI) Data Science

D.

Oracle Open Data

Question 35

As a data scientist, you require a pipeline to train ML models. When can a pipeline run be initiated?

Options:

A.

Pipeline can be initiated once it is created.

B.

Pipeline can be initiated during the pipeline run state.

C.

Pipeline can be initiated after the active state.

D.

Pipeline can be initiated before the active state.

Question 36

Which Oracle Cloud Infrastructure (OCI) Data Science policy is invalid?

Options:

A.

Allow group DataScienceGroup to use virtual-network-family in compartment DataScience

B.

Allow group DataScienceGroup to use data-science-model-sessions in compartment DataScience

C.

Allow dynamic-group DataScienceDynamicGroup to manage data-science-projects in compartment DataScience

D.

Allow dynamic-group DataScienceDynamicGroup to manage data-science-family in compartment DataScience

Question 37

How are datasets exported in the OCI Data Labeling service?

Options:

A.

As a binary file

B.

As an XML file

C.

As a line-delimited JSON file

D.

As a CSV file

Question 38

Which of the following best describes the principal goal of data science?

Options:

A.

To collect and archive exhaustive datasets from various source systems for corporate record-keeping uses.

B.

To mine and analyze large amounts of data in order to uncover information that can be used for operational improvements and business gains.

C.

To collect and prepare data for use as part of analytics applications.

D.

Data science is focused on output of the analysis.

Question 39

You are building a model and need input that represents data as morning, afternoon, or evening. However, the data contains a timestamp. What part of the Data Science lifecycle would you be in when creating the new variable?

Options:

A.

Model type selection

B.

Model validation

C.

Data access

D.

Feature engineering

Question 40

Which of the following TWO non-open source JupyterLab extensions has Oracle Cloud Infrastructure (OCI) Data Science developed and added to the notebook session experience?

Options:

A.

Environment Explorer

B.

Table of Contents

C.

Command Palette

D.

Notebook Examples

E.

Terminal

Question 41

Using Oracle AutoML, you are tuning hyperparameters on a supported model class and have specified a time budget. AutoML terminates computation once the time budget is exhausted. What would you expect AutoML to return in case the time budget is exhausted before hyperparameter tuning is completed?

Options:

A.

The current best-known hyperparameter configuration

B.

The last generated hyperparameter configuration

C.

A hyperparameter configuration with a minimum learning rate

D.

A random hyperparameter configuration

Question 42

As a data scientist, you use the Oracle Cloud Infrastructure (OCI) Language service to train custommodels. Which types of custom models can be trained?

Options:

A.

Image classification, Named Entity Recognition (NER)

B.

Text classification, Named Entity Recognition (NER)

C.

Sentiment Analysis, Named Entity Recognition (NER)

D.

Object detection, Text classification

Question 43

You are a researcher who requires access to large datasets. Which OCI service would you use?

Options:

A.

Oracle Databases

B.

ADW (Autonomous Data Warehouse)

C.

OCI Data Science

D.

Oracle Open Data

Question 44

You are using Oracle Cloud Infrastructure (OCI) Anomaly Detection to train a model to detect anomalies in pump sensor data. How does the required False Alarm Probability setting affect an anomaly detection model?

Options:

A.

It is used to disable the reporting of false alarms

B.

It changes the sensitivity of the model to detecting anomalies

C.

It determines how many false alarms occur before an error message is generated

D.

It adds a score to each signal indicating the probability that it’s a false alarm

Question 45

You are working as a data scientist for a healthcare company. They decide to analyze the data to find patterns in a large volume of electronic medical records. You are asked to build a PySpark solution to analyze these records in a JupyterLab notebook. What is the order of recommended stepsto develop a PySpark application in Oracle Cloud Infrastructure (OCI) Data Science?

Options:

A.

Install a Spark conda environment, configure core-site.xml, launch a notebook session, create a Data Flow application with the Accelerated Data Science (ADS) SDK, develop your PySpark application

B.

Configure core-site.xml, install a PySpark conda environment, create a Data Flow application with the Accelerated Data Science (ADS) SDK, develop your PySpark application, launch a notebook session

C.

Launch a notebook session, configure core-site.xml, install a PySpark conda environment, develop your PySpark application, create a Data Flow application with the Accelerated Data Science (ADS) SDK

D.

Launch a notebook session, install a PySpark conda environment, configure core-site.xml, develop your PySpark application, create a Data Flow application with the Accelerated Data Science (ADS) SDK

Question 46

You have been given a collection of digital files required for a business audit. They consist of several different formats that you would like to annotate using Oracle Cloud Infrastructure (OCI) Data Labeling. Which THREE types of files could this tool annotate?

Options:

A.

Video footage of a conversation in a conference room

B.

Images of computer server racks

C.

A typewritten document that details an annual budget

D.

A collection of purchase orders for office supplies

E.

An audio recording of a phone conversation

Question 47

You are preparing a configuration object necessary to create a Data Flow application. Which THREE parameter values should you provide?

Options:

A.

The path to the archive.zip file

B.

The local path to your PySpark script

C.

The compartment of the Data Flow application

D.

The bucket used to read/write the PySpark script in Object Storage

E.

The display name of the application

Page: 1 / 16
Total 158 questions