You are preparing a configuration object necessary to create a Data Flow application. Which THREE parameter values should you provide?
A. The path to the arhive.zip file.
B. The local path to your pySpark script.
C. The compartment of the Data Flow application.
D. The bucket used to read/write the pySpark script in Object Storage.
E. The display name of the application.
You are working as a data scientist for a healthcare company. They decide to analyze the data to find patterns in a large volume of electronic medical records. You are asked to build a PySpark solution to analyze these records in a JupyterLab notebook. What is the order of recommended steps to develop a PySpark application in Oracle Cloud Infrastructure (OCI) Data Science?
A. Launch a notebook session. Configure core-site.xml. Install a PySPark conda environ- ment. B. Develop your PySpark application Create a Data Flow application with the Ac- celerated Data Science (ADS) SOK
B. Configure core-site.xml. Install a PySPark conda environment. Create a Data Flow application with the Accelerated Data Science (ADS) SDK Develop your PySpark ap- plication. Launch a notebook session.
C. Launch a notebook session. Install a PySpark conda environment. Configure coresite.xml.
D. Develop your PySpark application. Create a Data Flow application with the Ac-celerated Data science (ADS) SDK.
E. Install a spark conda environment. Configure core-site.xml. Launch a notebook session:Create a Data Flow application with the Accelerated Data Science (ADS) SOK. Develop your PySpark application
What preparation steps are required to access an Oracle AI service SDK from a Data Science notebook session?
A. Call the Accented Data Science (ADS) command to enable Al integration
B. Create and upload the API signing key and config file
C. Import the REST API
D. Create and upload execute.py and runtime.yaml
Which Oracle Accelerated Data Science (ADS) classes can be used for easy access to data sets from reference libraries and index websites, such as scikit-learn?
A. ADSTurner
B. DatasetFactory
C. SecretKeeper
D. Dataset Browser
You have created a model, and you want to use the Accelerated Data Science (ADS) SDK to deploy this model. Where can you save the artifacts to deploy this model with ADS?
A. Model Depository
B. Model Catalog
C. OCI Vault
D. Data Science Artifactory
You are a computer vision engineer building an image recognition model. You decide to use Oracle Data Labeling to annotate your image data. Which of the following THREE are possible ways to annotate an image in Data Labeling?
A. Adding labels to image using semantic segmentation, by drawing multiple bounding boxes to an image.
B. Adding a single label to an image.
C. Adding labels to an image by drawing bounding box to an image, is not supported by Data Labeling
D. Adding labels to an image using object detection, by drawing bounding boxes to an im- age.
E. Adding multiple labels to an image.
You have built a machine model to predict whether a bank customer is going to default on a loan. You want to use Local Interpretable Model-Agnostic Explanations (LIME) to understand a specific prediction. What is the key idea behind LIME?
A. Model-agnostic techniques are more interpretable than techniques that are dependent on the types of models.
B. Local explanation techniques are model agnostic, while global explanation techniques are not.
C. Global behavior of a machine learning model may be complex, while the local behavior may be approximated with a simpler surrogate model.
D. Global and local behaviors of machine learning models are similar.
You are a data scientist working inside a notebook session and you attempt to pip install a package from a public repository that is not included in your condo environment. After running this command, you get a network timeout error. What might be missing from your networking configuration?
A. Service Gateway with private subnet access.
B. NAT Gateway with public internet access.
C. FastConnect to an on-premises network.
D. Primary Virtual Network Interface Card (VNIC).
You have a complex Python code project that could benefit from using Data Science Jobs as it is a repeatable machine learning model training task. The project contains many subfolder and classes. What is the best way to run this project as a job?
A. ZIP the entire code project folder, upload it as a Job artifact on job creation and set JOB_RUN_ENTRYPOINT to point to the main executable file.
B. ZIP the entire code project folder and upload it as a Job artifact on job creation, Jobs identities the main executable file automatically.
C. Rewrite your code so that a single executable Python or Bash/Shell script file.
D. ZIP the entire code project folder and upload it as a Job artifact Jobs automatically identifies That main top level where the code is run.
You are using a third-party Continuous Integration/Continuous Delivery (CI/CD) tool to create a pipeline for preparing and training models. How would you integrate a third-party tool outside Oracle Cloud Infrastructure (OCI) to access Data Science Jobs?
A. Third-party software can access Data Science Jobs by using any of the OCI Software Development Kits (SDKs).
B. Data Science Jobs does not accept code from third-party tools, therefore you need to run the pipeline externally.
C. Third-party tools use authentication keys to create and run.
D. Data Science Jobs Data Science Jobs is not accessible from outside OCI.