Skip to content

Commit fed6358

Browse files
committed
added docs
1 parent cf24b43 commit fed6358

File tree

1 file changed

+27
-0
lines changed

1 file changed

+27
-0
lines changed

ads/feature_store/docs/source/feature_group.rst

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -122,6 +122,33 @@ The ``.materialise()`` method takes the following parameter:
122122
.. seealso::
123123
Refer :ref:`Data types` supported by feature store
124124

125+
126+
Materialise Stream
127+
==================
128+
You can call the ``materialise_stream() -> FeatureGroupJob`` method of the ``FeatureGroup`` instance to load the streaming data to feature group. To persist the feature_group and save feature_group data along the metadata in the feature store, call the ``materialise_stream()``
129+
130+
The ``.materialise_stream()`` method takes the following parameter:
131+
- ``input_dataframe``: Features in Streaming Dataframe to be saved.
132+
- ``query_name``: It is possible to optionally specify a name for the query to make it easier to recognise in the Spark UI. Defaults to ``None``.
133+
- ``ingestion_mode``: Specifies how data of a streaming DataFrame/Dataset is written to a streaming sink.
134+
- ``"append"``: Only the new rows in the streaming DataFrame/Dataset will be written to the sink. If the query doesn’t contain aggregations, it will be equivalent to
135+
- append mode. Defaults to ``"append"``.
136+
- ``"complete"``: All the rows in the streaming DataFrame/Dataset will be written to the sink every time there is some update.
137+
- ``"update"``: only the rows that were updated in the streaming DataFrame/Dataset will be written to the sink every time there are some updates.
138+
- ``await_termination``: Waits for the termination of this query, either by query.stop() or by an exception. If the query has terminated with an exception, then the exception will be thrown. If timeout is set, it returns whether the query has terminated or not within the timeout seconds. Defaults to ``False``.
139+
- ``timeout``: Only relevant in combination with ``await_termination=True``.
140+
- Defaults to ``None``.
141+
- ``checkpoint_dir``: Checkpoint directory location. This will be used to as a reference to from where to resume the streaming job. If ``None`` then hsfs will construct as "insert_stream_" + online_topic_name. Defaults to ``None``.
142+
- ``write_options``: Additional write options for Spark as key-value pairs.
143+
- Defaults to ``{}``.
144+
145+
.. seealso::
146+
:ref:`Feature Group Job`
147+
148+
.. seealso::
149+
Refer :ref:`Data types` supported by feature store
150+
151+
125152
Delete
126153
======
127154

0 commit comments

Comments
 (0)