You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/user_guide/apachespark/dataflow-spark-magic.rst
+7-7Lines changed: 7 additions & 7 deletions
Original file line number
Diff line number
Diff line change
@@ -23,8 +23,8 @@ This section demonstrates how to run interactive Spark workloads on a long lasti
23
23
24
24
**Notebook Examples:**
25
25
26
-
* `Introduction to the Oracle Cloud Infrastructure Data Flow Studio <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/master/notebook_examples/pyspark-data_flow_studio-introduction.ipynb>`__
27
-
* `Spark NLP within Oracle Cloud Infrastructure Data Flow Studio <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/master/notebook_examples/pyspark-data_flow_studio-spark_nlp.ipynb>`__
26
+
* `Introduction to the Oracle Cloud Infrastructure Data Flow Studio <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/pyspark-data_flow_studio-introduction.ipynb>`_
27
+
* `Spark NLP within Oracle Cloud Infrastructure Data Flow Studio <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/pyspark-data_flow_studio-spark_nlp.ipynb>`_
28
28
29
29
Prerequisite
30
30
============
@@ -72,7 +72,7 @@ Data Flow Spark Magic is used for interactively working with remote Spark cluste
72
72
73
73
%load_ext dataflow.magics
74
74
75
-
Use the `%help` method to get a list of all the available commands, along with a list of their arguments and example calls.
75
+
Use the ``%help`` method to get a list of all the available commands, along with a list of their arguments and example calls.
76
76
77
77
.. code-block:: python
78
78
@@ -130,7 +130,7 @@ To help you save resources and reduce time on management, Spark `dynamic allocat
130
130
131
131
**Example command with third-party libraries**
132
132
133
-
The Data Flow Sessions support `custom dependencies <https://docs.oracle.com/iaas/data-flow/using/third-party-libraries.htm>`__ in the form of Python wheels or virtual environments. You might want to make native code or other assets available within your Spark runtime. The dependencies can be attached by using the `archiveUri` attribute.
133
+
The Data Flow Sessions support `custom dependencies <https://docs.oracle.com/iaas/data-flow/using/third-party-libraries.htm>`__ in the form of Python wheels or virtual environments. You might want to make native code or other assets available within your Spark runtime. The dependencies can be attached by using the ``archiveUri`` attribute.
134
134
135
135
.. code-block:: python
136
136
@@ -148,7 +148,7 @@ The Data Flow Sessions support `custom dependencies <https://docs.oracle.com/iaa
148
148
149
149
**Example command with the Data Catalog Hive Metastore**
150
150
151
-
The `Data Catalog Hive Metastore <https://docs.oracle.com/iaas/data-catalog/using/metastore.htm>`__ provides schema definitions for objects in structured and unstructured data assets. Use the `metastoreId` to access the Data Catalog Metastore.
151
+
The `Data Catalog Hive Metastore <https://docs.oracle.com/iaas/data-catalog/using/metastore.htm>`__ provides schema definitions for objects in structured and unstructured data assets. Use the ``metastoreId`` to access the Data Catalog Metastore.
152
152
153
153
.. code-block:: python
154
154
@@ -195,7 +195,7 @@ Example path : ``oci://<your-bucket>@<your-tenancy-namespace>/conda_environments
195
195
196
196
.. versionadded:: 2.8.7
197
197
198
-
The `Data Flow Pools <https://docs.oracle.com/en-us/iaas/data-flow/using/pools.htm>`__ achieve fast job startup, resource isolation, manage budgets, and prioritize your Spark workloads. Use the `poolId` to use the Pool resources.
198
+
The `Data Flow Pools <https://docs.oracle.com/en-us/iaas/data-flow/using/pools.htm>`__ achieve fast job startup, resource isolation, manage budgets, and prioritize your Spark workloads. Use the ``poolId`` to use the Pool resources.
199
199
200
200
.. code-block:: python
201
201
@@ -256,7 +256,7 @@ To re-activate the existing session, use the ``%activate_session`` magic command
256
256
257
257
Use Existing Session
258
258
********************
259
-
To connect to the existing session use the `%use_session` magic command.
259
+
To connect to the existing session use the ``%use_session`` magic command.
Copy file name to clipboardExpand all lines: docs/source/user_guide/configuration/authentication.rst
+1-2Lines changed: 1 addition & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -27,7 +27,7 @@ Within your notebook session, you can choose to use the resource principal to au
27
27
API Keys
28
28
========
29
29
30
-
This is the default method of authentication. You can also authenticate as your own personal IAM user by creating or uploading OCI configuration and API key files inside your notebook session environment. The OCI configuration file contains the necessary credentials to authenticate your user against the model catalog and other OCI services like Object Storage. The example notebook, `api_keys.ipynb` demonstrates how to create these files.
30
+
This is the default method of authentication. You can also authenticate as your own personal IAM user by creating or uploading OCI configuration and API key files inside your notebook session environment. The OCI configuration file contains the necessary credentials to authenticate your user against the model catalog and other OCI services like Object Storage. To setup API Key refer to `Required Keys and OCIDs <https://docs.oracle.com/en-us/iaas/Content/API/Concepts/apisigningkey.htm>`_ and `SDK and CLI Configuration File <https://docs.oracle.com/en-us/iaas/Content/API/Concepts/sdkconfig.htm>`_.
31
31
32
32
The ``getting-started.ipynb`` notebook in the home directory of the notebook session environment demonstrates all the steps needed to create the configuration file and the keys. Follow the steps in that notebook before importing and using ADS in your notebooks.
33
33
@@ -79,4 +79,3 @@ In the this example, the default authentication uses API keys specified with the
79
79
os_auth = authutil.resource_principal() # use resource principal to as the preferred way to access object store
80
80
81
81
``AuthContext`` context class can also be used to specify the desired type of authentication. It supports API key configuration, resource principal, and instance principal authentication, as well as predefined signers, callable signers, or API keys configurations from specified locations. See `API Documentation <../../ads.common.html#ads.common.auth.AuthContext>`__ for more details.
7. To find the location of the ``sqlnet.ora`` file, the ``TNS_ADMIN`` environment variable must point to that location. We suggest that you create a Python dictionary to store all of the connection information. In this example, this dictionary is called ``creds``. It is generally poor security practice to store credentials in your notebook. We recommend that you use the ``ads-examples/ADB_working_with.ipynb`` notebook example that demonstrates how to store them outside the notebook in a configuration file.
36
+
7. To find the location of the ``sqlnet.ora`` file, the ``TNS_ADMIN`` environment variable must point to that location. We suggest that you create a Python dictionary to store all of the connection information. In this example, this dictionary is called ``creds``. It is generally poor security practice to store credentials in your notebook. We recommend that you use the `Bank Graph Example Notebook <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/graph_insight-autonomous_database.ipynb>`_ notebook example that demonstrates how to store them outside the notebook in a configuration file.
37
37
38
38
The environment variable should be set in your notebooks. For example:
39
39
@@ -72,4 +72,3 @@ Messages similar to the following display if the connection is successful:
72
72
:align:center
73
73
74
74
An introduction to loading data from ADB into ADS using ``cx_Oracle`` and ``SQLAlchemy`` is in :ref:`Loading Data <loading-data-10>`.
This is the default method of authentication. You can also authenticate as your own personal IAM user by creating or uploading OCI configuration and API key files inside your notebook session environment. The OCI configuration file contains the necessary credentials to authenticate your user against the model catalog and other OCI services like Object Storage. The example notebook, `api_keys.ipynb` demonstrates how to create these files.
35
+
This is the default method of authentication. You can also authenticate as your own personal IAM user by creating or uploading OCI configuration and API key files inside your notebook session environment. The OCI configuration file contains the necessary credentials to authenticate your user against the model catalog and other OCI services like Object Storage. To setup API Key refer to `Required Keys and OCIDs <https://docs.oracle.com/en-us/iaas/Content/API/Concepts/apisigningkey.htm>`_ and `SDK and CLI Configuration File <https://docs.oracle.com/en-us/iaas/Content/API/Concepts/sdkconfig.htm>`_.
36
36
37
37
The ``getting-started.ipynb`` notebook in the home directory of the notebook session environment demonstrates all the steps needed to create the configuration file and the keys. Follow the steps in that notebook before importing and using ADS in your notebooks.
38
38
@@ -117,7 +117,7 @@ Use these steps to access Oracle ADB:
7. To find the location of the ``sqlnet.ora`` file, the ``TNS_ADMIN`` environment variable must point to that location. We suggest that you create a Python dictionary to store all of the connection information. In this example, this dictionary is called ``creds``. It is generally poor security practice to store credentials in your notebook. We recommend that you use the ``ads-examples/ADB_working_with.ipynb`` notebook example that demonstrates how to store them outside the notebook in a configuration file.
120
+
7. To find the location of the ``sqlnet.ora`` file, the ``TNS_ADMIN`` environment variable must point to that location. We suggest that you create a Python dictionary to store all of the connection information. In this example, this dictionary is called ``creds``. It is generally poor security practice to store credentials in your notebook. We recommend that you use the `Bank Graph Example Notebook <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/graph_insight-autonomous_database.ipynb>`_ notebook example that demonstrates how to store them outside the notebook in a configuration file.
121
121
122
122
The environment variable should be set in your notebooks. For example:
123
123
@@ -440,11 +440,11 @@ dictionary. To store your secret, just modify the dictionary.
440
440
'password': 'MySecretPassword',
441
441
'database_type': 'oracle'}
442
442
443
-
Note, to connect to an Oracle database the `database_name` value should be its
443
+
Note, to connect to an Oracle database the ``database_name`` value should be its
444
444
connection identifier. You can find the connection identifier by extracting the
445
-
credential wallet zip file and opening the `tnsnames.ora` file
445
+
credential wallet zip file and opening the ``tnsnames.ora`` file
446
446
(connection_identifier = (...)). Usually the connection identifier will
447
-
end with `_high`, `_medium` or `_low` i.e. `'MyDatabaseName_high'`.
447
+
end with ``_high``, ``_medium`` or ``_low`` i.e. ``'MyDatabaseName_high'``.
Copy file name to clipboardExpand all lines: docs/source/user_guide/data_flow/dataflow.rst
+4-6Lines changed: 4 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -791,7 +791,7 @@ You can work with different compression formats within Data Flow. For example, s
791
791
792
792
Other compression formats that Data Flow supports include snappy Parquet, and Gzip on both CSV and Parquet.
793
793
794
-
You might have query that you want to run in Data Flow from previous explorations, review the `dataflow.ipynb` notebook example that shows you how to submit a job to Data Flow.
794
+
You might have query that you want to run in Data Flow from previous explorations.
795
795
796
796
.. code-block:: python3
797
797
@@ -1025,12 +1025,11 @@ The required variables to set up are:
1025
1025
that the secret is the credential needed to access a database. This
1026
1026
notebook is designed so that any secret can be stored as long as it
1027
1027
is in the form of a dictionary. To store your secret, just modify the
1028
-
dictionary, see the ``vault.ipynb`` example notebook for detailed
1029
-
steps to generate this OCID.
1028
+
dictionary.
1030
1029
2. ``tnsname``: A TNS name valid for the database.
1031
1030
3. ``wallet_path``: The local path to your wallet ZIP file, see the
1032
-
``autonomous_database.ipynb`` example notebook for instructions on
1033
-
accessing the wallet file.
1031
+
`Loading Data from Different Sources Using Pandas and Dask <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/load_data-object_storage-hive-autonomous-database.ipynb>`_
1032
+
example notebook for instructions on accessing the wallet file.
1034
1033
1035
1034
.. code-block:: python3
1036
1035
@@ -1137,4 +1136,3 @@ This notebook creates a table in your database with the name specified with ``ta
1137
1136
)
1138
1137
else:
1139
1138
print("Skipping as it appears that you do not have tnsname specified.")
Copy file name to clipboardExpand all lines: docs/source/user_guide/model_training/ads_tuner.rst
+3-3Lines changed: 3 additions & 3 deletions
Original file line number
Diff line number
Diff line change
@@ -341,7 +341,7 @@ space is being used for your model class when ``strategy`` is
341
341
``perfunctory`` or ``detailed`` use the ``search_space()`` method to see
342
342
the details.
343
343
344
-
The ``adstuner_search_space_update.ipynb`` notebook has detailed
344
+
The `Introduction to ADSTuner <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/hyperparameter_tuning.ipynb>`_ notebook has detailed
345
345
examples about how to work with and update the search space.
346
346
347
347
The next cell displaces the search strategy and the search space.
@@ -363,7 +363,7 @@ parameter. When it is set to ``False``, the tuning process runs
363
363
asynchronously so it runs in the background and allows you to continue
364
364
your work in the notebook. When ``synchronous`` is set to ``True``, the
365
365
notebook is blocked until ``tune()`` finishes running. The
366
-
``adntuner_sync_and_async.ipynb`` notebook illustrates this feature in a
366
+
`Introduction to ADSTuner <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/hyperparameter_tuning.ipynb>`_ notebook illustrates this feature in a
367
367
more detailed way.
368
368
369
369
The ``ADSTuner`` object needs to know when to stop tuning. The
@@ -739,7 +739,7 @@ You can change the search space in the following three ways:
739
739
hyperparameter or make any changes to a hyperparameter that is based on
740
740
a categorical distribution. You need to initiate a new ``ADSTuner``
741
741
object for those cases. For more detailed information, review the
742
-
``adstuner_search_space_update.ipynb`` notebook.
742
+
`Introduction to ADSTuner <https://github.com/oracle-samples/oci-data-science-ai-samples/blob/main/notebook_examples/hyperparameter_tuning.ipynb>`_ notebook.
743
743
744
744
The next cell switches to a ``detailed`` strategy. All previous values
745
745
set for ``C``, ``solver``, and ``max_iter`` are kept, and ``ADSTuner``
0 commit comments