-
Notifications
You must be signed in to change notification settings - Fork 90
sql query
Up until now we have focused on doing processing and setting up alerts without having to write any code. That is powerful and will satisfy a bunch of scenarios. However, when the no code experience isn't sufficient to meet the needs, you can use the power of Spark SQL for more complex processing, and do so without having to learn Scala. In this tutorial you will learn how to use SQL for processing data, chain queries, output data to sinks.
-
Open a Flow or create a new one using first tutorial to create a pipeline in 5 minutes
-
Switch to the Query tab. This is where you can write all your SQL code. Data Accelerator supports full Spark SQL, and enhances it with few more keywords that you will learn about in future tutorials. For now, write a statement such as:
--DataXQuery-- T1 = SELECT * FROM DataXProcessedInput;
- Each query is separated by comment "--DataXQuery--" and ends with a semi-colon. You can chain queries.
--DataXQuery-- DeviceWindowedInput = SELECT deviceDetails.deviceId, deviceDetails.deviceType, eventTimeStamp, deviceDetails.homeId, deviceDetails.status FROM T1 GROUP BY deviceId, deviceType, eventTimeStamp, homeId, status;
--DataXQuery-- DeviceInfo = SELECT deviceId, deviceType, homeId, MAX(eventTimeStamp) AS MaxEventTime, MIN(status) AS MinReading, MAX(status) AS MaxReading FROM DeviceWindowedInput GROUP BY deviceId, deviceType, homeId;
OUTPUT DeviceInfo TO myCosmosDB;