Monday, March 27, 2023
HomeBig DataJob Notifications in SQL Stream Builder

Job Notifications in SQL Stream Builder

Particular co-author credit: Adam Andras Toth, Software Engineer Intern

With enterprises’ wants for information analytics and processing getting extra complicated by the day, Cloudera goals to maintain up with these wants, providing continuously evolving, cutting-edge options to all of your information associated issues. Cloudera Stream Processing goals to take real-time information analytics to the following stage. We’re excited to spotlight job monitoring with notifications, a brand new function for SQL Stream Builder (SSB).

The sudden failing of a posh information pipeline can result in devastating penaltiesparticularly if it goes unnoticed. An actual-time monetary fraud detector, or a posh structure gathering and aggregating information to create insights and permit their prospects to make data-driven selectionsthese are techniques which have little to no room for error or prolonged downtime. This is the reason we construct job notifications performance into SSB, to ship most reliability in your complicated real-time information pipelines.

Job notifications will help just be sure you can detect failed jobs with out checking on the UI, which may save lots of time for the consumer. This function could be very helpful, particularly when the consumer has quite a few jobs working and retaining monitor of their state can be exhausting with out notifications.

First, we wish to introduce the structure of job notifications. Allow us to use a determine to exhibit how job notifications match into SSB, then we are going to focus on every kind individually.


In SSB you may handle a number of tasks. Initiatives symbolize the software program improvement life cycle (SDLC) in SQL Stream Builder (SSB): you may create an setting for creating SQL jobs, verifying the outcomes after which pushing them to manufacturing. The assets for job creation are shared throughout the jobs of a venture, and there are assets out there that can be utilized between the tasks. The fundamental idea of a venture can be expanded for collaboration by sharing the tasks with crew members in streaming SQL console, or utilizing supply management to synchronize the venture with a Git repository. 

Job notifications additionally belong to tasks. Meaning in a single venture you may outline a number of notifications, and people notifications can solely belong to the roles of that venture. Within the determine under, you may see the structure of a venture from the angle of job notifications. As of now there are two sorts of notifications: electronic mail and webhook. The notifications can be organized into teams. The good thing about that is that if you wish to assign the identical set of notifications to a number of jobs you don’t have to do that one after the other in each job, you may simply create a notification group and assign that to the roles. One notification might be included in a number of teams and a bunch may even comprise one other group.

Within the determine under, the identical job notifications are marked with the identical colour. As you may see within the venture we’ve got three jobs. Within the first one we solely have notifications, so if that job fails these 4 notifications will hearth. In the second we’ve got a webhook notification and a notification group that has one other webhook and an electronic mail notification, so if this job fails these three notifications will go off. The third job has a webhook notification, a bunch that accommodates an electronic mail notification, and one other notification group that has two notifications, so if this job fails these 4 notifications will hearth.


As I discussed earlier than, there are two sorts of notifications and you may assign them to teams. I’ll first introduce placeholders, which you should utilize to create notifications.


The e-mail message or webhook request that’s despatched upon the set off for a notification might be fully custom-made. Greater than that, SSB additionally permits the utilization of placeholders, which can be utilized to offer all obligatory info within the notification. With the power to customise messages and to make use of placeholders, customers can even probably be capable of robotically parse the incoming notifications and create automated responses for them, thus guaranteeing that essential pipelines might be restarted with out requiring human intervention. 

The placeholders presently out there for utilization are:

  • jobName
  • jobStatus
  • jobStatusDescription
  • ssbJobId
  • flinkJobId
  • clusterId
  • lastException

You should use a placeholder within the following format: “Houston we’ve got an issue, your job with identify ${jobName} has failed.”

E mail notifications

E mail notifications are (as you possibly can guess from its identify) sending emails to the given electronic mail deal with upon job failure. To make this work some CM properties must be configured:

  • Mail server host for job notifications: The host of the SMTP server for job failure notifications
  • Mail server username for job notifications: The username to entry the SMTP server for job failure notifications
  • Mail server password for job notifications: The password to entry the SMTP server for job failure notifications
  • SMTP authentication for job notifications: Allow SMTP authentication for job notifications (default worth: True)
  • StartTLS for job notifications: Use the StartTLS command to determine a safe connection to the SMTP server for job notifications (default worth: True)
  • Job notifications sender mail deal with: Sender mail deal with for job notifications
  • Mail server port for job notifications: The port of the SMTP server for job failure notifications (default worth: 587)

You probably have these items arrange correctly and also you add a notification to your job, it’s best to get an electronic mail if the job fails.

Webhook notifications

With webhook notifications you may make webhook requests upon a job failure. In the event you use the placeholders accurately, then you should utilize the outlined webhook endpoints of exterior purposes to deal with the failures in a extra environment friendly approach. (For instance, you may arrange a webhook notification with Slack to ship you a message instantly if a job fails.)

Within the case of webhook notifications you may set one property in CM:

  • Job notifications webhook sender parallelism: Variety of threads utilized by the job notification job to name user-specified webhooks when notifying a couple of failed or lacking job (default worth: 10)

DISCLAIMER: The payload template of a webhook notification should be a legitimate JSON! Additionally be certain that to place placeholders inside quotes!


  • “identify”: ${jobName} is invalid
  • “identify”:”${jobName}” is legitimate
  • “identify”:”no matter i would like right here ${jobName}” can also be legitimate

Notification teams

As I discussed above you may assign your notifications into teams. This fashion you don’t want so as to add all of the notifications to the roles one after the other. A cool factor in regards to the teams is that they will additionally comprise different notification teams. 

SSB’s job notifications function is a cool method to hold monitor of your failing jobs and thus decrease the downtime of them. You simply want to ensure the “allow job notifications” performance in CM is checked. The job-monitoring job periodically queries the state of your jobs, and triggers the assigned notifications if a failed job is discovered. The examine interval in CM might be configured with the job notifications monitoring interval property (default worth: 60s).

On this part I’ll present you some video examples for the usages of the job notifications.

Create and use an E mail notification:

Create and use a Webhook notification:

Create and use a Notification Group

Anyone can check out SSB utilizing the Stream Processing Neighborhood Version (CSP-CE). CE makes creating stream processors simple, as it may be completed proper out of your desktop or another improvement node. Analysts, information scientists, and builders can now consider new options, develop SQL-based stream processors domestically utilizing SQL Stream Builder powered by Flink, and develop Kafka Customers/Producers and Kafka Join Connectors, all domestically earlier than shifting to manufacturing in CDP.



Please enter your comment!
Please enter your name here

Most Popular

Recent Comments