Skip to content

Commit 37250d2

Browse files
authored
Merge branch 'main' into ODSC-37153/action_to_publish_docs
2 parents 682bd94 + 5738f9f commit 37250d2

File tree

28 files changed

+106
-71
lines changed

28 files changed

+106
-71
lines changed

.github/workflows/publish-to-readthedocs.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,7 @@ on:
1313
- main
1414
paths:
1515
- 'docs/**'
16+
1617
env:
1718
RTDS_ADS_PROJECT: https://readthedocs.org/api/v3/projects/accelerated-data-science
1819
RTDS_ADS_TOKEN: ${{ secrets.RTDS_ADS_TOKEN }}

README.md

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,21 +1,21 @@
1-
# Oracle Accelerated Data Science SDK (ADS)
1+
# Oracle Accelerated Data Science (ADS)
2+
3+
[![PyPI](https://img.shields.io/pypi/v/oracle-ads.svg?style=for-the-badge&logo=pypi&logoColor=white)](https://pypi.org/project/oracle-ads/) [![Python](https://img.shields.io/pypi/pyversions/oracle-ads.svg?style=for-the-badge&logo=pypi&logoColor=white)](https://pypi.org/project/oracle-ads/) [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg?style=for-the-badge&logo=pypi&logoColor=white)](https://github.com/ambv/black)
24

3-
[![PyPI](https://img.shields.io/pypi/v/oracle-ads.svg?style=for-the-badge&logo=pypi&logoColor=white)](https://pypi.org/project/oracle-ads/) [![Python](https://img.shields.io/pypi/pyversions/oracle-ads.svg?style=for-the-badge&logo=pypi&logoColor=white)](https://pypi.org/project/oracle-ads/)
45

56
The [Oracle Accelerated Data Science (ADS) SDK](https://accelerated-data-science.readthedocs.io/en/latest/index.html) is maintained by the Oracle Cloud Infrastructure (OCI) [Data Science service](https://docs.oracle.com/en-us/iaas/data-science/using/data-science.htm) team. It speeds up common data science activities by providing tools that automate and simplify common data science tasks. Additionally, provides data scientists a friendly pythonic interface to OCI services. Some of the more notable services are OCI Data Science, Model Catalog, Model Deployment, Jobs, ML Pipelines, Data Flow, Object Storage, Vault, Big Data Service, Data Catalog, and the Autonomous Database. ADS gives you an interface to manage the life cycle of machine learning models, from data acquisition to model evaluation, interpretation, and model deployment.
67

78
With ADS you can:
89

9-
- Read datasets from Oracle Object Storage, Oracle RDBMS (ATP/ADW/On-prem), AWS S3 and other sources into `Pandas dataframes`.
10-
- Use feature types to characterize your data, create meaning summary statistics and plot. Use the warning and validation system to test the quality of your data.
11-
- Tune models using hyperparameter optimization with the `ADSTuner` tool.
12-
- Generate detailed evaluation reports of your model candidates with the `ADSEvaluator` module.
13-
- Save machine learning models to the [OCI Data Science Model Catalog](https://docs.oracle.com/en-us/iaas/data-science/using/models-about.htm).
14-
- Deploy models as HTTP endpoints with [Model Deployment](https://docs.oracle.com/en-us/iaas/data-science/using/model-dep-about.htm).
15-
- Launch distributed ETL, data processing, and model training jobs in Spark with [OCI Data Flow](https://docs.oracle.com/en-us/iaas/data-flow/using/home.htm).
16-
- Train machine learning models in OCI Data Science [Jobs](https://docs.oracle.com/en-us/iaas/data-science/using/jobs-about.htm).
17-
- Define and run an end-to-end machine learning orchestration covering all the steps of machine learning lifecycle in a repeatable, continuous [ML Pipelines](https://accelerated-data-science.readthedocs.io/en/latest/user_guide/pipeline/overview.html#).
18-
- Manage the life cycle of conda environments through the `ads conda` command line interface (CLI).
10+
- Read datasets from Oracle Object Storage, Oracle RDBMS (ATP/ADW/On-prem), AWS S3 and other sources into `Pandas dataframes`.
11+
- Tune models using hyperparameter optimization with the `ADSTuner` tool.
12+
- Generate detailed evaluation reports of your model candidates with the `ADSEvaluator` module.
13+
- Save machine learning models to the [OCI Data Science Model Catalog](https://docs.oracle.com/en-us/iaas/data-science/using/models-about.htm).
14+
- Deploy models as HTTP endpoints with [Model Deployment](https://docs.oracle.com/en-us/iaas/data-science/using/model-dep-about.htm).
15+
- Launch distributed ETL, data processing, and model training jobs in Spark with [OCI Data Flow](https://docs.oracle.com/en-us/iaas/data-flow/using/home.htm).
16+
- Train machine learning models in OCI Data Science [Jobs](https://docs.oracle.com/en-us/iaas/data-science/using/jobs-about.htm).
17+
- Define and run an end-to-end machine learning orchestration covering all the steps of machine learning lifecycle in a repeatable, continuous [ML Pipelines](https://accelerated-data-science.readthedocs.io/en/latest/user_guide/pipeline/overview.html#).
18+
- Manage the life cycle of conda environments through the `ads conda` command line interface (CLI).
1919

2020
## Installation
2121

docs/requirements.txt

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,10 @@ autodoc
22
nbsphinx
33
sphinx
44
sphinxcontrib-napoleon
5-
sphinx-rtd-theme
65
sphinx_copybutton
76
sphinx_code_tabs
87
sphinx-autobuild
9-
sphinx-design
8+
sphinx-autorun
109
oracle_ads
1110
furo
1211
IPython
52.3 KB
Loading
8 KB
Loading

docs/source/_static/oracle_logo.png

-12.6 KB
Binary file not shown.

docs/source/conf.py

Lines changed: 24 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -23,21 +23,20 @@
2323
version = release = __import__("ads").__version__
2424

2525
extensions = [
26-
"sphinx_rtd_theme",
2726
"sphinx.ext.napoleon",
2827
"sphinx.ext.autodoc",
2928
"sphinx.ext.doctest",
30-
"sphinx.ext.todo",
3129
"sphinx.ext.mathjax",
3230
"sphinx.ext.ifconfig",
33-
"sphinx.ext.graphviz",
34-
"sphinx.ext.inheritance_diagram",
31+
"sphinx.ext.autodoc",
3532
"sphinx.ext.todo",
36-
"sphinx.ext.viewcode",
33+
"sphinx.ext.extlinks",
34+
"sphinx.ext.intersphinx",
35+
"sphinx.ext.graphviz",
3736
"nbsphinx",
3837
"sphinx_code_tabs",
39-
"sphinx_design",
40-
"sphinx_copybutton"
38+
"sphinx_copybutton",
39+
"sphinx_autorun",
4140
]
4241

4342
# Add any paths that contain templates here, relative to this directory.
@@ -63,16 +62,29 @@
6362
# directories to ignore when looking for source files.
6463
# This pattern also affects html_static_path and html_extra_path.
6564
# exclude_patterns = []
66-
exclude_patterns = ['build', '**.ipynb_checkpoints']
65+
exclude_patterns = ['build', '**.ipynb_checkpoints', 'Thumbs.db', '.DS_Store']
6766

68-
# The name of the Pygments (syntax highlighting) style to use.
69-
pygments_style = None
7067
language = "en"
7168

7269
html_theme = "furo"
73-
html_logo = "_static/oracle_logo.png"
7470
html_static_path = ["_static"]
75-
html_css_files = ["pied-piper-admonition.css"]
71+
72+
html_title = f"{project} v{release}"
73+
74+
# Disable the generation of the various indexes
75+
html_use_modindex = False
76+
html_use_index = False
77+
78+
# html_css_files = [
79+
# 'pied-piper-admonition.css',
80+
# ]
81+
82+
html_theme_options = {
83+
"light_logo": "logo-light-mode.png",
84+
"dark_logo": "logo-dark-mode.png",
85+
}
86+
87+
7688
htmlhelp_basename = "pydoc"
7789

7890

docs/source/index.rst

Lines changed: 30 additions & 31 deletions
Original file line numberDiff line numberDiff line change
@@ -6,8 +6,8 @@
66
library and CLI for Machine learning engineers to work with Cloud Infrastructure (CPU and GPU VMs, Storage etc, Spark) for Data, Models,
77
Notebooks, Pipelines and Jobs.
88

9-
Oracle Accelerated Data Science SDK (ADS)
10-
=========================================
9+
Oracle Accelerated Data Science (ADS)
10+
=====================================
1111
|PyPI|_ |Python|_ |Notebook Examples|_
1212

1313
.. |PyPI| image:: https://img.shields.io/pypi/v/oracle-ads.svg?style=for-the-badge&logo=pypi&logoColor=white
@@ -66,47 +66,44 @@ Oracle Accelerated Data Science SDK (ADS)
6666

6767
modules
6868

69-
.. admonition:: Oracle Accelerated Data Science (ADS) SDK
69+
.. admonition:: Oracle Accelerated Data Science (ADS)
70+
:class: note
7071

71-
The Oracle Accelerated Data Science (ADS) SDK is maintained by the Oracle Cloud Infrastructure Data Science service team. It speeds up common data science activities by providing tools that automate and/or simplify common data science tasks, along with providing a data scientist friendly pythonic interface to Oracle Cloud Infrastructure (OCI) services, most notably OCI Data Science, Data Flow, Object Storage, and the Autonomous Database. ADS gives you an interface to manage the lifecycle of machine learning models, from data acquisition to model evaluation, interpretation, and model deployment.
72+
Oracle Accelerated Data Science (ADS) is maintained by the Oracle Cloud Infrastructure Data Science service team. It speeds up common data science activities by providing tools that automate and/or simplify common data science tasks, along with providing a data scientist friendly pythonic interface to Oracle Cloud Infrastructure (OCI) services, most notably OCI Data Science, Data Flow, Object Storage, and the Autonomous Database. ADS gives you an interface to manage the lifecycle of machine learning models, from data acquisition to model evaluation, interpretation, and model deployment.
7273

73-
With ADS you can:
74+
With ADS you can:
7475

75-
- Read datasets from Oracle Object Storage, Oracle RDBMS (ATP/ADW/On-prem), AWS S3, and other sources into Pandas dataframes.
76-
- Easily compute summary statistics on your dataframes and perform data profiling.
77-
- Tune models using hyperparameter optimization with the ADSTuner tool.
78-
- Generate detailed evaluation reports of your model candidates with the ADSEvaluator module.
79-
- Save machine learning models to the OCI Data Science Models.
80-
- Deploy those models as HTTPS endpoints with Model Deployment.
81-
- Launch distributed ETL, data processing, and model training jobs in Spark with OCI Data Flow.
82-
- Train machine learning models in OCI Data Science Jobs.
83-
- Manage the lifecycle of conda environments through the ads conda command line interface (CLI).
84-
- Distributed Training with PyTorch, Horovod and Dask
76+
- Read datasets from Oracle Object Storage, Oracle RDBMS (ATP/ADW/On-prem), AWS S3, and other sources into Pandas dataframes.
77+
- Easily compute summary statistics on your dataframes and perform data profiling.
78+
- Tune models using hyperparameter optimization with the ADSTuner tool.
79+
- Generate detailed evaluation reports of your model candidates with the ADSEvaluator module.
80+
- Save machine learning models to the OCI Data Science Models.
81+
- Deploy those models as HTTPS endpoints with Model Deployment.
82+
- Launch distributed ETL, data processing, and model training jobs in Spark with OCI Data Flow.
83+
- Train machine learning models in OCI Data Science Jobs.
84+
- Manage the lifecycle of conda environments through the ads conda command line interface (CLI).
85+
- Distributed Training with PyTorch, Horovod and Dask
8586

8687

8788
.. admonition:: Installation
89+
:class: note
8890

8991
python3 -m pip install oracle-ads
9092

9193

9294
.. admonition:: Source Code
95+
:class: note
9396

9497
`https://github.com/oracle/accelerated-data-science <https://github.com/oracle/accelerated-data-science>`_
9598

96-
.. code:: ipython3
97-
99+
.. code-block:: python3
98100
>>> import ads
99101
>>> ads.hello()
100102
101-
O o-o o-o
102-
/ \ | \ |
103-
o---o| O o-o
104-
| || / |
105-
o oo-o o--o
103+
.. runblock:: pycon
106104

107-
ADS SDK version: X.Y.Z
108-
Pandas version: x.y.z
109-
Debug mode: False
105+
>>> import ads
106+
>>> ads.hello()
110107

111108

112109
Additional Documentation
@@ -115,6 +112,8 @@ Additional Documentation
115112
- `OCI Data Science and AI services Examples <https://github.com/oracle/oci-data-science-ai-samples>`_
116113
- `Oracle AI & Data Science Blog <https://blogs.oracle.com/ai-and-datascience/>`_
117114
- `OCI Documentation <https://docs.oracle.com/en-us/iaas/data-science/using/data-science.htm>`_
115+
- `OCIFS Documentation <https://ocifs.readthedocs.io/en/latest/>`_
116+
- `Example Notebooks <https://github.com/oracle-samples/oci-data-science-ai-samples/tree/master/notebook_examples>`_
118117

119118
Examples
120119
++++++++
@@ -147,25 +146,25 @@ This example uses SQL injection safe binding variables.
147146

148147
.. code-block:: python3
149148
150-
import ads
151-
import pandas as pd
149+
import ads
150+
import pandas as pd
152151
153-
connection_parameters = {
152+
connection_parameters = {
154153
"user_name": "<user_name>",
155154
"password": "<password>",
156155
"service_name": "<tns_name>",
157156
"wallet_location": "<file_path>",
158-
}
157+
}
159158
160-
df = pd.DataFrame.ads.read_sql(
159+
df = pd.DataFrame.ads.read_sql(
161160
"""
162161
SELECT *
163162
FROM SH.SALES
164163
WHERE ROWNUM <= :max_rows
165164
""",
166165
bind_variables={ max_rows : 100 },
167166
connection_parameters=connection_parameters,
168-
)
167+
)
169168
170169
More Examples
171170
~~~~~~~~~~~~~

docs/source/user_guide/apachespark/dataflow-spark-magic.rst

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -79,6 +79,7 @@ Use the `%help` method to get a list of all the available commands, along with a
7979
%help
8080
8181
.. admonition:: Tip
82+
:class: note
8283

8384
To access the docstrings of any magic command and figure out what arguments to provide, simply add ``?`` at the end of the command. For instance: ``%create_session?``
8485

docs/source/user_guide/apachespark/dataflow.rst

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -41,6 +41,7 @@ Define config. If you have not yet configured your dataflow setting, or would li
4141
Use the config defined above to submit the cell.
4242

4343
.. admonition:: Tip
44+
:class: note
4445

4546
Get more information about the dataflow extension by running ``%dataflow -h``
4647

@@ -131,11 +132,13 @@ To submit your notebook to DataFlow using the ``ads`` CLI, run:
131132
ads opctl run -s <folder where notebook is located> -e <notebook name> -b dataflow
132133
133134
.. admonition:: Tip
135+
:class: note
134136

135137
You can avoid running cells that are not DataFlow environment compatible by tagging the cells and then providing the tag names to ignore. In the following example cells that are tagged ``ignore`` and ``remove`` will be ignored -
136138
``--exclude-tag ignore --exclude-tag remove``
137139

138140
.. admonition:: Tip
141+
:class: note
139142

140143
You can run the notebook in your local pyspark environment before submitting to ``DataFlow`` using the same CLI with ``-b local``
141144

0 commit comments

Comments
 (0)