Skip to content

Commit d746582

Browse files
authored
chore(demo/nifi-kafka-druid-earthquake-data demo): Update screenshots and docs (#198)
chore: update screenshots and docs for nifi-kafka-druid-earthquake-data demo
1 parent f70ee02 commit d746582

File tree

7 files changed

+11
-9
lines changed

7 files changed

+11
-9
lines changed
Loading
Loading
Loading
Loading
Loading
Loading

docs/modules/demos/pages/nifi-kafka-druid-earthquake-data.adoc

Lines changed: 11 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -209,25 +209,26 @@ Log in with the username `admin` and password `adminadmin`.
209209

210210
image::nifi-kafka-druid-earthquake-data/nifi_2.png[]
211211

212-
You can see the started ProcessGroup consisting of two processors.
212+
You can see the started ProcessGroup consisting of three processors.
213213
The first one - `InvokeHTTP`, fetches the CSV file from the Internet and puts it into the queue of the next processor.
214-
The second processor - `PublishKafkaRecord_2_6`, parses the CSV file, converts it to JSON records and writes them out into Kafka.
214+
The second processor - `SplitRecords`, takes the single FlowFile (NiFi Record) which contains all CSV records and splits it into chunks of 2000 records, which are then separately put into the queue of the next processor.
215+
The third one - `PublishKafkaRecord`, parses the CSV chunk, converts it to JSON records and writes them out into Kafka.
215216

216217
Double-click on the `InvokeHTTP` processor to show the processor details.
217218

218219
image::nifi-kafka-druid-earthquake-data/nifi_3.png[]
219220

220-
Head over to the tab `PROPERTIES`.
221+
Head over to the `Properties` tab.
221222

222223
image::nifi-kafka-druid-earthquake-data/nifi_4.png[]
223224

224225
Here, you can see the setting `HTTP URL`, which specifies the download URL from where the CSV file is retrieved.
225-
Close the processor details popup by clicking `OK`.
226-
Afterwards, double-click on the processor `PublishKafkaRecord_2_6`.
226+
Close the processor details popup by clicking `Close`.
227+
Afterwards, double-click on the processor `PublishKafkaRecord`.
227228

228229
image::nifi-kafka-druid-earthquake-data/nifi_5.png[]
229230

230-
The Kafka connection details within this processor - like broker addresses and topic names - are specified.
231+
The Kafka connection service, which contains the connection details, and the topic name is specified.
231232
It uses the `CSVReader` to parse the downloaded CSV and the `JsonRecordSetWriter` to split it into individual JSON records before writing it out.
232233

233234
== Druid
@@ -247,12 +248,13 @@ By clicking on `Supervisors` at the top, you can see the running ingestion jobs.
247248
image::nifi-kafka-druid-earthquake-data/druid_2.png[]
248249

249250
You can see additional information after clicking on the magnification glass to the right side of the `RUNNING` supervisor.
250-
On the tab `Statistics` on the left, you can see the number of processed records as well as the number of errors.
251+
On the tab `Task stats` on the left, you can see the number of processed records as well as the number of errors.
251252

252253
image::nifi-kafka-druid-earthquake-data/druid_3.png[]
253254

254-
The statistics show that Druid ingested `5074` records during the last minute and has ingested 3 million records already.
255-
All entries have been consumed successfully, indicated by having no `processWithError`, `thrownAway` or `unparseable` records.
255+
The statistics show that Druid ingested `13279` records per second within the last minute and has ingested around 600,000 records already.
256+
All entries have been consumed successfully, indicated by having no `processWithError`, `thrownAway` or `unparseable` records in the output of the `View raw`
257+
button at the top right.
256258

257259
=== Query the Data Source
258260

0 commit comments

Comments
 (0)