package streaming
- Alphabetic
- Public
- All
Type Members
-
final
class
DataStreamReader extends Logging
Interface used to load a streaming
Dataset
from external storage systems (e.g.Interface used to load a streaming
Dataset
from external storage systems (e.g. file systems, key-value stores, etc). UseSparkSession.readStream
to access this.- Annotations
- @Evolving()
- Since
2.0.0
-
final
class
DataStreamWriter[T] extends AnyRef
Interface used to write a streaming
Dataset
to external storage systems (e.g.Interface used to write a streaming
Dataset
to external storage systems (e.g. file systems, key-value stores, etc). UseDataset.writeStream
to access this.- Annotations
- @Evolving()
- Since
2.0.0
-
trait
GroupState[S] extends LogicalGroupState[S]
:: Experimental ::
:: Experimental ::
Wrapper class for interacting with per-group state data in
mapGroupsWithState
andflatMapGroupsWithState
operations onKeyValueGroupedDataset
.Detail description on
[map/flatMap]GroupsWithState
operation -------------------------------------------------------------- Both,mapGroupsWithState
andflatMapGroupsWithState
inKeyValueGroupedDataset
will invoke the user-given function on each group (defined by the grouping function inDataset.groupByKey()
) while maintaining user-defined per-group state between invocations. For a static batch Dataset, the function will be invoked once per group. For a streaming Dataset, the function will be invoked for each group repeatedly in every trigger. That is, in every batch of theStreamingQuery
, the function will be invoked once for each group that has data in the trigger. Furthermore, if timeout is set, then the function will invoked on timed out groups (more detail below).The function is invoked with following parameters.
- The key of the group.
- An iterator containing all the values for this group.
- A user-defined state object set by previous invocations of the given function.
In case of a batch Dataset, there is only one invocation and state object will be empty as there is no prior state. Essentially, for batch Datasets,
[map/flatMap]GroupsWithState
is equivalent to[map/flatMap]Groups
and any updates to the state and/or timeouts have no effect.The major difference between
mapGroupsWithState
andflatMapGroupsWithState
is that the former allows the function to return one and only one record, whereas the latter allows the function to return any number of records (including no records). Furthermore, theflatMapGroupsWithState
is associated with an operation output mode, which can be eitherAppend
orUpdate
. Semantically, this defines whether the output records of one trigger is effectively replacing the previously output records (from previous triggers) or is appending to the list of previously output records. Essentially, this defines how the Result Table (refer to the semantics in the programming guide) is updated, and allows us to reason about the semantics of later operations.Important points to note about the function (both mapGroupsWithState and flatMapGroupsWithState).
- In a trigger, the function will be called only the groups present in the batch. So do not assume that the function will be called in every trigger for every group that has state.
- There is no guaranteed ordering of values in the iterator in the function, neither with batch, nor with streaming Datasets.
- All the data will be shuffled before applying the function.
- If timeout is set, then the function will also be called with no values.
See more details on
GroupStateTimeout
below.
Important points to note about using
GroupState
.- The value of the state cannot be null. So updating state with null will throw
IllegalArgumentException
. - Operations on
GroupState
are not thread-safe. This is to avoid memory barriers. - If
remove()
is called, thenexists()
will returnfalse
,get()
will throwNoSuchElementException
andgetOption()
will returnNone
- After that, if
update(newState)
is called, thenexists()
will again returntrue
,get()
andgetOption()
will return the updated value.
Important points to note about using
GroupStateTimeout
.- The timeout type is a global param across all the groups (set as
timeout
param in[map|flatMap]GroupsWithState
, but the exact timeout duration/timestamp is configurable per group by callingsetTimeout...()
inGroupState
. - Timeouts can be either based on processing time (i.e.
GroupStateTimeout.ProcessingTimeTimeout
) or event time (i.e.GroupStateTimeout.EventTimeTimeout
). - With
ProcessingTimeTimeout
, the timeout duration can be set by callingGroupState.setTimeoutDuration
. The timeout will occur when the clock has advanced by the set duration. Guarantees provided by this timeout with a duration of D ms are as follows:- Timeout will never be occur before the clock time has advanced by D ms
- Timeout will occur eventually when there is a trigger in the query (i.e. after D ms). So there is a no strict upper bound on when the timeout would occur. For example, the trigger interval of the query will affect when the timeout actually occurs. If there is no data in the stream (for any group) for a while, then their will not be any trigger and timeout function call will not occur until there is data.
- Since the processing time timeout is based on the clock time, it is affected by the variations in the system clock (i.e. time zone changes, clock skew, etc.).
- With
EventTimeTimeout
, the user also has to specify the the the event time watermark in the query usingDataset.withWatermark()
. With this setting, data that is older than the watermark are filtered out. The timeout can be set for a group by setting a timeout timestamp usingGroupState.setTimeoutTimestamp()
, and the timeout would occur when the watermark advances beyond the set timestamp. You can control the timeout delay by two parameters - (i) watermark delay and an additional duration beyond the timestamp in the event (which is guaranteed to be newer than watermark due to the filtering). Guarantees provided by this timeout are as follows:- Timeout will never be occur before watermark has exceeded the set timeout.
- Similar to processing time timeouts, there is a no strict upper bound on the delay when the timeout actually occurs. The watermark can advance only when there is data in the stream, and the event time of the data has actually advanced.
- When the timeout occurs for a group, the function is called for that group with no values, and
GroupState.hasTimedOut()
set to true. - The timeout is reset every time the function is called on a group, that is, when the group has new data, or the group has timed out. So the user has to set the timeout duration every time the function is called, otherwise there will not be any timeout set.
Scala example of using GroupState in
mapGroupsWithState
:// A mapping function that maintains an integer state for string keys and returns a string. // Additionally, it sets a timeout to remove the state if it has not received data for an hour. def mappingFunction(key: String, value: Iterator[Int], state: GroupState[Int]): String = { if (state.hasTimedOut) { // If called when timing out, remove the state state.remove() } else if (state.exists) { // If state exists, use it for processing val existingState = state.get // Get the existing state val shouldRemove = ... // Decide whether to remove the state if (shouldRemove) { state.remove() // Remove the state } else { val newState = ... state.update(newState) // Set the new state state.setTimeoutDuration("1 hour") // Set the timeout } } else { val initialState = ... state.update(initialState) // Set the initial state state.setTimeoutDuration("1 hour") // Set the timeout } ... // return something } dataset .groupByKey(...) .mapGroupsWithState(GroupStateTimeout.ProcessingTimeTimeout)(mappingFunction)
Java example of using
GroupState
:// A mapping function that maintains an integer state for string keys and returns a string. // Additionally, it sets a timeout to remove the state if it has not received data for an hour. MapGroupsWithStateFunction<String, Integer, Integer, String> mappingFunction = new MapGroupsWithStateFunction<String, Integer, Integer, String>() { @Override public String call(String key, Iterator<Integer> value, GroupState<Integer> state) { if (state.hasTimedOut()) { // If called when timing out, remove the state state.remove(); } else if (state.exists()) { // If state exists, use it for processing int existingState = state.get(); // Get the existing state boolean shouldRemove = ...; // Decide whether to remove the state if (shouldRemove) { state.remove(); // Remove the state } else { int newState = ...; state.update(newState); // Set the new state state.setTimeoutDuration("1 hour"); // Set the timeout } } else { int initialState = ...; // Set the initial state state.update(initialState); state.setTimeoutDuration("1 hour"); // Set the timeout } ... // return something } }; dataset .groupByKey(...) .mapGroupsWithState( mappingFunction, Encoders.INT, Encoders.STRING, GroupStateTimeout.ProcessingTimeTimeout);
- S
User-defined type of the state to be stored for each group. Must be encodable into Spark SQL types (see
Encoder
for more details).
- Annotations
- @Experimental() @Evolving()
- Since
2.2.0
- class GroupStateTimeout extends AnyRef
- class OutputMode extends AnyRef
-
class
SinkProgress extends Serializable
Information about progress made for a sink in the execution of a StreamingQuery during a trigger.
Information about progress made for a sink in the execution of a StreamingQuery during a trigger. See StreamingQueryProgress for more information.
- Annotations
- @Evolving()
- Since
2.1.0
-
class
SourceProgress extends Serializable
Information about progress made for a source in the execution of a StreamingQuery during a trigger.
Information about progress made for a source in the execution of a StreamingQuery during a trigger. See StreamingQueryProgress for more information.
- Annotations
- @Evolving()
- Since
2.1.0
-
class
StateOperatorProgress extends Serializable
Information about updates made to stateful operators in a StreamingQuery during a trigger.
Information about updates made to stateful operators in a StreamingQuery during a trigger.
- Annotations
- @Evolving()
-
trait
StreamingQuery extends AnyRef
A handle to a query that is executing continuously in the background as new data arrives.
A handle to a query that is executing continuously in the background as new data arrives. All these methods are thread-safe.
- Annotations
- @Evolving()
- Since
2.0.0
-
class
StreamingQueryException extends Exception
Exception that stopped a StreamingQuery.
Exception that stopped a StreamingQuery. Use
cause
get the actual exception that caused the failure.- Annotations
- @Evolving()
- Since
2.0.0
-
abstract
class
StreamingQueryListener extends AnyRef
Interface for listening to events related to StreamingQueries.
Interface for listening to events related to StreamingQueries.
- Annotations
- @Evolving()
- Since
2.0.0
- Note
The methods are not thread-safe as they may be called from different threads.
-
class
StreamingQueryManager extends Logging
A class to manage all the StreamingQuery active in a
SparkSession
.A class to manage all the StreamingQuery active in a
SparkSession
.- Annotations
- @Evolving()
- Since
2.0.0
-
class
StreamingQueryProgress extends Serializable
Information about progress made in the execution of a StreamingQuery during a trigger.
Information about progress made in the execution of a StreamingQuery during a trigger. Each event relates to processing done for a single trigger of the streaming query. Events are emitted even when no new data is available to be processed.
- Annotations
- @Evolving()
- Since
2.1.0
-
class
StreamingQueryStatus extends Serializable
Reports information about the instantaneous status of a streaming query.
Reports information about the instantaneous status of a streaming query.
- Annotations
- @Evolving()
- Since
2.1.0
- class Trigger extends AnyRef
-
case class
ProcessingTime(intervalMs: Long) extends Trigger with Product with Serializable
A trigger that runs a query periodically based on the processing time.
A trigger that runs a query periodically based on the processing time. If
interval
is 0, the query will run as fast as possible.Scala Example:
df.writeStream.trigger(ProcessingTime("10 seconds")) import scala.concurrent.duration._ df.writeStream.trigger(ProcessingTime(10.seconds))
Java Example:
df.writeStream.trigger(ProcessingTime.create("10 seconds")) import java.util.concurrent.TimeUnit df.writeStream.trigger(ProcessingTime.create(10, TimeUnit.SECONDS))
- Annotations
- @Evolving() @deprecated
- Deprecated
(Since version 2.2.0) use Trigger.ProcessingTime(intervalMs)
- Since
2.0.0
Value Members
-
object
StreamingQueryListener
Companion object of StreamingQueryListener that defines the listener events.
Companion object of StreamingQueryListener that defines the listener events.
- Annotations
- @Evolving()
- Since
2.0.0
Deprecated Value Members
-
object
ProcessingTime extends Serializable
Used to create ProcessingTime triggers for StreamingQuerys.
Used to create ProcessingTime triggers for StreamingQuerys.
- Annotations
- @Evolving() @deprecated
- Deprecated
(Since version 2.2.0) use Trigger.ProcessingTime(intervalMs)
- Since
2.0.0