The matters lined are:
- Information pipeline structure
- Excessive-scale knowledge ingestion
- Information transformation and processing
- Information storage
- Staging knowledge supply
- Operational knowledge
- Fingers-on train
Picture supply
Information Pipeline Structure
An information pipeline structure consists of elements and techniques that present providers at each layer of the end-to-end knowledge stream. The phases are knowledge ingestion, knowledge transformation and processing, knowledge storage, knowledge supply, and knowledge consumption for functions and operational knowledge shops.
The info stream begins with an arbitrary knowledge supply which is usually a database, flat information, enterprise functions, real-time streaming knowledge, and so forth.
Excessive Scale Information Ingestion
Information ingestion is the method of buying, mapping, and validating the acquired knowledge and transferring the information to the information processing part. Structure sides related to the information ingestion course of for the trendy knowledge pipeline structure are as follows:
- Information mapping: That is the method of mapping the information construction to the pipeline format. Map the information utilizing configuration or flatten the information on the fly. A schema-agnostic method is elastic. Any format corresponding to JSON, XML, or CSV may be flattened to an unstructured format that the pipeline processes.
- Information elasticity: Unflatten the unstructured pipeline knowledge right into a structured format for evaluation.
- Excessive-scale ingestion: Capacity to deal with the amount and velocity of knowledge being generated
- Batch-based ingestion: Pulls knowledge from the supply on a scheduled foundation
- On-time ingestion: Receives knowledge as a real-time stream
- Ordering: An information occasion is ordered based mostly on the timestamp of the receipt.
- Information consistency: No duplicates and no knowledge loss; identical knowledge occasion however nonetheless one copy; as soon as acquired saved in a log
Information Transformation and Processing
This part leads to the cleansing and transformation of knowledge for the downstream system, adopted by processing for the following knowledge storage part.
Structure sides related to the information transformation/processing part for the trendy knowledge pipeline structure:
- Cleansing: Make the information related.
- Transformation: Change values for relevance to the use case.
- Processing at scale: Add many parallel knowledge pipelines. Information ought to keep constant.
- Processing with accuracy: No duplicates on repetitive knowledge occasions; the supply dataset should match the vacation spot dataset.
Information Storage
This part persists the processed knowledge right into a storage system like an information warehouse or a datalake. The info saved in these techniques can be utilized for a lot of use instances corresponding to analytics, coaching AI fashions, bridging knowledge silos between enterprise functions, enterprise intelligence, and so forth.
Structure sides related to the information storage course of are as follows:
- Storing knowledge: Retailer knowledge for future retrieval, and evaluation, or every other use case.
- Organizing knowledge: Organizing knowledge optimized for question efficiency
- Safety and compliance: Information lifecycle for retention and regulatory compliance if relevant
Staging Information Supply
This part is the place knowledge pipelines can differ of their implementations. Some could name the earlier knowledge storage part the ultimate vacation spot. Some present this part to ship processed knowledge to particular person storage techniques used for operational techniques. Nevertheless, that leads to two phases: staging the information adopted by an automatic course of or a guide course of to make the information operational. Additionally they differ in what storage techniques they help out of the field. These techniques are normal techniques like relational databases, NoSQL, real-time search techniques, and so forth.
Structure sides related to the staging knowledge supply part embody:
- Parallel, or serial: Information processing may be parallel, as one route shops knowledge to a datalake or knowledge warehouse, and one other delivers knowledge for staging. Parallel is environment friendly and sooner.
- Information connector help: An information pipeline implementation will present help for traditional storage techniques like MongoDB, MySQL, ElasticSearch, Snowflake, and so forth.
- Information connector framework: It might additionally present an information connector framework to hook up with customized storage techniques.
Supporting this part with an information pipeline has the next benefits:
- Enterprise knowledge silos: Join a number of storage techniques inside your enterprise to maintain them in sync in real-time with knowledge consistency.
- Supporting exterior knowledge companions: Ship knowledge to a number of exterior storage techniques.
Operational Information Part
That is additionally a part the place knowledge pipelines can differ of their implementation. That is the part that processes staged knowledge in operational techniques and makes it stay knowledge.
Structure sides related to the operational knowledge part are:
- Automated and real-time: An information pipeline can present an information integration framework that may invoke customized logic to course of staged knowledge and make it stay knowledge as a part of the information stream course of.
- Batch or guide: On this case, a company can use a background course of or manually course of the staged knowledge to make it stay knowledge.
Hand-On Train
This hands-on train will contain implementing an end-to-end knowledge pipeline. The implementation will contain the information stream that can cowl the next phases:
- Information ingestion
- Information transformation and processing
- Information storage
- Staging knowledge supply
It makes use of the open-source Braineous knowledge platform (obtain right here). Ignore the “Too big to load” message and click on on the Obtain button.
- This tutorial is situated beneath braineous-1.0.0-cr3/tutorials/get-started.
Set up
Observe that Braineous has the next service dependencies:
The most recent launch, CR3, requires adjusting these providers to your localhost setup for these providers. The following launch, CR4, will streamline the localhost service detection and a Kubernetes container out of the field.
Listed below are the localhost directions for every service earlier than operating Braineous on localhost:
Zookeeper
Modify braineous-1.0.0-cr3/bin/start_zookeeper.sh
script from:
~/mumma/braineous/infrastructure/kafka_2.13-3.5.0/bin/zookeeper-server-start.sh
/Customers/babyboy/mumma/braineous/infrastructure/kafka_2.13-3.5.0/config/zookeeper.properties
. . . to your native Zookeeper bin and config directories.
Kafka
Modify braineous-1.0.0-cr3/bin/start_kafka.sh
script from:
~/mumma/braineous/infrastructure/kafka_2.13-3.5.0/bin/kafka-server-start.sh
/Customers/babyboy/mumma/braineous/infrastructure/kafka_2.13-3.5.0/config/server.properties
. . . to your native Kafka bin and config directories.
Flink
Modify braineous-1.0.0-cr3/bin/start_flink.sh
script from:
~/mumma/braineous/infrastructure/flink-1.18.1/bin/start-cluster.sh
. . . to your native Flink bin listing.
Hive Metastore
Modify braineous-1.0.0-cr3/bin/start_metastore.sh
script from:
export HADOOP_HOME=~/mumma/braineous/infrastructure/hadoop-3.3.6
export HIVE_HOME=~/mumma/braineous/infrastructure/apache-hive-3.1.3-bin
export HADOOP_CLASSPATH=$HIVE_HOME/conf:$HIVE_HOME/lib
~/mumma/braineous/infrastructure/apache-hive-3.1.3-bin/bin/hive --service metastore
. . . to your respective Hadoop/Hive native installations. As well as, change the next property within the local_hive_installation/conf/hive-site.xml file from my localhost file system location to yours.
hive.metastore.warehouse.dir
/Customers/babyboy/data_warehouse/hive
location of default database for the warehouse
Braineous
- Modify braineous-1.0.0-cr3/conf/braineous.config:
#hive
hive_conf_directory=/Customers/babyboy/mumma/braineous/infrastructure/apache-hive-3.1.3-bin/conf
table_directory=file:///Customers/babyboy/datalake/
- Modify the
hive_conf_directory
andtable_directory
properties to level to directories in your localhost filesystem. - CR4 will make the localhost out-of-the-box expertise optimum, and likewise present a Kubernetes Container.
As soon as your localhost providers and configuration for Zookeeper, Kafka, Flink, and Hive Metastore are arrange, observe the remainder of the steps for the hands-on train to create the information pipeline.
Step 1: Set up MongoDB
brew set up mongodb-community@7.0
Step 2: Begin Zookeeper
cd braineous-1.0.0-cr3/bin
Step 3: Begin Kafka
cd braineous-1.0.0-cr3/bin
Step 4: Begin Flink
cd braineous-1.0.0-cr3/bin
Step 5: Begin Hive Metastore
cd braineous-1.0.0-cr3/bin
Step 6: Begin Braineous
cd braineous-1.0.0-cr3/bin
Step 7: Examine Braineous Set up Occasion Success
cd braineous-1.0.0-cr3/bin
Step 8: Create a Tenant with an API Key and Secret
cd braineous-1.0.0-cr3/bin
Please use this API key and API secret for the tutorial. Extra directions within the knowledge ingestion part to observe.
Information Pipeline Registration
Register an information pipe with the information ingestion engine:
{
"pipeId": "yyya",
"entity": "abc",
"configuration": [
{
"stagingStore" : "com.appgallabs.dataplatform.targetSystem.core.driver.MongoDBStagingStore",
"name": "yyya",
"config": {
"connectionString": "mongodb://localhost:27017",
"database": "yyya",
"collection": "data",
"jsonpathExpressions": []
}
}
]
}
For particulars concerning the pipe configuration please consult with the Developer Documentation Information.
Supply Information
[
{
"id" : 1,
"name": "name_1",
"age": 46,
"addr": {
"email": "name_1@email.com",
"phone": "123"
}
},
{
"id": "2",
"name": "name_2",
"age": 55,
"addr": {
"email": "name_2@email.com",
"phone": "1234"
}
}
]
A dataset may be loaded from any knowledge supply corresponding to a database, legacy manufacturing knowledge retailer, stay knowledge feed, third-party knowledge supply, Kafka stream, and so forth. On this instance, the dataset is loaded from a classpath useful resource situated at src/important/sources/dataset/knowledge.json.
Information Ingestion
Information Ingestion Shopper Implementation
DataPlatformService dataPlatformService = DataPlatformService.getInstance();
String apiKey = "ffb2969c-5182-454f-9a0b-f3f2fb0ebf75";
String apiSecret = "5960253b-6645-41bf-b520-eede5754196e";
String datasetLocation = "dataset/data.json";
String json = Util.loadResource(datasetLocation);
JsonElement datasetElement = JsonUtil.validateJson(json);
System.out.println("*****DATA_SET******");
JsonUtil.printStdOut(datasetElement);
String configLocation = "pipe_config/pipe_config.json";
String pipeConfigJson = Util.loadResource(configLocation);
JsonObject configJson = JsonUtil.validateJson(pipeConfigJson).getAsJsonObject();
String pipeId = configJson.get("pipeId").getAsString();
String entity = configJson.get("entity").getAsString();
System.out.println("*****PIPE_CONFIGURATION******");
JsonUtil.printStdOut(configJson);
//configure the DataPipeline Shopper
Configuration configuration = new Configuration().
ingestionHostUrl("http://localhost:8080/").
apiKey(apiKey).
apiSecret(apiSecret).
streamSizeInObjects(0);
dataPlatformService.configure(configuration);
//register pipe
dataPlatformService.registerPipe(configJson);
System.out.println("*****PIPE_REGISTRATION_SUCCESS******");
//ship supply knowledge by way of the pipeline
dataPlatformService.sendData(pipeId, entity,datasetElement.toString());
System.out.println("*****DATA_INGESTION_SUCCESS******");
Please use the API key and API secret generated in Step 8 of the Set up part for the variables:
Information Ingestion Shopper Output
Anticipated output from the ingestion consumer:
[INFO] ------------------------------------------------------------------------
SLF4J: Did not load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for additional particulars.
*****DATA_SET******
******ARRAY_SIZE: 2**********
[
{
"id": 1,
"name": "name_1",
"age": 46,
"addr": {
"email": "name_1@email.com",
"phone": "123"
}
},
{
"id": "2",
"name": "name_2",
"age": 55,
"addr": {
"email": "name_2@email.com",
"phone": "1234"
}
}
]
**********************
*****PIPE_CONFIGURATION******
{
"pipeId": "yyya",
"entity": "abc",
"configuration": [
{
"stagingStore": "com.appgallabs.dataplatform.targetSystem.core.driver.MongoDBStagingStore",
"name": "yyya",
"config": {
"connectionString": "mongodb://localhost:27017",
"database": "yyya",
"collection": "data",
"jsonpathExpressions": []
}
}
]
}
**********************
*****PIPE_REGISTRATION_SUCCESS******
***SENDING_DATA_START*****
*****DATA_INGESTION_SUCCESS******
Information Pipeline Server Output
Anticipated output on the information pipeline server:
2024-08-05 19:06:09,071 INFO [org.ehc.siz.imp.JvmInformation] (pool-7-thread-1) Detected JVM knowledge mannequin settings of: 64-Bit OpenJDK JVM with Compressed OOPs
2024-08-05 19:06:09,178 INFO [org.ehc.siz.imp.AgentLoader] (pool-7-thread-1) Failed to connect to VM and cargo the agent: class java.io.IOException: Cannot connect to present VM
WARNING: An unlawful reflective entry operation has occurred
WARNING: Unlawful reflective entry by org.ehcache.sizeof.ObjectGraphWalker (file:/Customers/babyboy/mumma/braineous/data_platform/cr3/braineous_dataplatform/releases/braineous-1.0.0-cr3/bin/dataplatform-1.0.0-cr3-runner.jar) to subject java.util.LinkedList.first
WARNING: Please take into account reporting this to the maintainers of org.ehcache.sizeof.ObjectGraphWalker
WARNING: Use --illegal-access=warn to allow warnings of additional unlawful reflective entry operations
WARNING: All unlawful entry operations can be denied in a future releaseMONGODB: DATA_STORED_SUCCESSFULLY
2024-08-05 19:06:16,562 INFO [org.apa.had.hiv.con.HiveConf] (pool-7-thread-1) Discovered configuration file null
2024-08-05 19:06:17,398 WARN [org.apa.had.uti.NativeCodeLoader] (pool-7-thread-1) Unable to load native-hadoop library in your platform... utilizing builtin-java lessons the place relevant
2024-08-05 19:06:17,492 INFO [org.apa.fli.tab.cat.hiv.HiveCatalog] (pool-7-thread-1) Setting hive conf dir as /Customers/babyboy/mumma/braineous/infrastructure/apache-hive-3.1.3-bin/conf
2024-08-05 19:06:17,791 INFO [org.apa.fli.tab.cat.hiv.HiveCatalog] (pool-7-thread-1) Created HiveCatalog 'ffbaaaacaaaaaaaaafaaaabafafafbaebfaa'
2024-08-05 19:06:17,871 INFO [org.apa.had.hiv.met.HiveMetaStoreClient] (pool-7-thread-1) Attempting to hook up with metastore with URI thrift://0.0.0.0:9083
2024-08-05 19:06:17,900 INFO [org.apa.had.hiv.met.HiveMetaStoreClient] (pool-7-thread-1) Opened a connection to metastore, present connections: 1
2024-08-05 19:06:18,639 INFO [org.apa.had.hiv.met.HiveMetaStoreClient] (pool-7-thread-1) Related to metastore.
2024-08-05 19:06:18,641 INFO [org.apa.had.hiv.met.RetryingMetaStoreClient] (pool-7-thread-1) RetryingMetaStoreClient proxy=class org.apache.hadoop.hive.metastore.HiveMetaStoreClient ugi=babyboy (auth:SIMPLE) retries=1 delay=1 lifetime=0
2024-08-05 19:06:18,907 INFO [org.apa.fli.tab.cat.hiv.HiveCatalog] (pool-7-thread-1) Related to Hive metastore
2024-08-05 19:06:19,725 INFO [org.apa.fli.tab.cat.CatalogManager] (pool-7-thread-1) Set the present default catalog as [ffbaaaacaaaaaaaaafaaaabafafafbaebfaa] and the present default database as [default].
2024-08-05 19:06:21,351 INFO [org.apa.fli.tab.cat.hiv.HiveCatalog] (pool-10-thread-1) Setting hive conf dir as /Customers/babyboy/mumma/braineous/infrastructure/apache-hive-3.1.3-bin/conf
2024-08-05 19:06:21,381 INFO [org.apa.fli.tab.cat.hiv.HiveCatalog] (pool-10-thread-1) Created HiveCatalog 'ffbaaaacaaaaaaaaafaaaabafafafbaebfaa'
2024-08-05 19:06:21,387 INFO [org.apa.had.hiv.met.HiveMetaStoreClient] (pool-10-thread-1) Attempting to hook up with metastore with URI thrift://0.0.0.0:9083
2024-08-05 19:06:21,388 INFO [org.apa.had.hiv.met.HiveMetaStoreClient] (pool-10-thread-1) Opened a connection to metastore, present connections: 2
2024-08-05 19:06:21,396 INFO [org.apa.had.hiv.met.HiveMetaStoreClient] (pool-10-thread-1) Related to metastore.
2024-08-05 19:06:21,397 INFO [org.apa.had.hiv.met.RetryingMetaStoreClient] (pool-10-thread-1) RetryingMetaStoreClient proxy=class org.apache.hadoop.hive.metastore.HiveMetaStoreClient ugi=babyboy (auth:SIMPLE) retries=1 delay=1 lifetime=0
2024-08-05 19:06:21,399 INFO [org.apa.fli.tab.cat.hiv.HiveCatalog] (pool-10-thread-1) Related to Hive metastore
2024-08-05 19:06:21,459 INFO [org.apa.fli.tab.cat.CatalogManager] (pool-10-thread-1) Set the present default catalog as [ffbaaaacaaaaaaaaafaaaabafafafbaebfaa] and the present default database as [default].
(
`id` STRING,
`title` STRING,
`age` STRING,
`addr.e mail` STRING,
`addr.cellphone` STRING,
`0` STRING
)
2024-08-05 19:06:21,863 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlCreateCatalog doesn't comprise a setter for subject catalogName
2024-08-05 19:06:21,864 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlCreateCatalog can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,866 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlCreateView doesn't comprise a setter for subject viewName
2024-08-05 19:06:21,866 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlCreateView can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,868 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlAlterViewRename doesn't comprise a getter for subject newViewIdentifier
2024-08-05 19:06:21,868 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlAlterViewRename doesn't comprise a setter for subject newViewIdentifier
2024-08-05 19:06:21,868 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlAlterViewRename can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,871 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlAlterViewProperties doesn't comprise a setter for subject propertyList
2024-08-05 19:06:21,871 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlAlterViewProperties can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,873 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlAlterViewAs doesn't comprise a setter for subject newQuery
2024-08-05 19:06:21,873 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlAlterViewAs can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,874 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlAddPartitions doesn't comprise a setter for subject ifPartitionNotExists
2024-08-05 19:06:21,874 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlAddPartitions can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,875 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlDropPartitions doesn't comprise a setter for subject ifExists
2024-08-05 19:06:21,875 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlDropPartitions can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,876 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dql.SqlShowPartitions doesn't comprise a getter for subject tableIdentifier
2024-08-05 19:06:21,877 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dql.SqlShowPartitions doesn't comprise a setter for subject tableIdentifier
2024-08-05 19:06:21,877 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.dql.SqlShowPartitions can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,878 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dml.SqlTruncateTable doesn't comprise a getter for subject tableNameIdentifier
2024-08-05 19:06:21,878 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dml.SqlTruncateTable doesn't comprise a setter for subject tableNameIdentifier
2024-08-05 19:06:21,879 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.dml.SqlTruncateTable can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,879 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dql.SqlShowFunctions doesn't comprise a setter for subject requireUser
2024-08-05 19:06:21,879 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.dql.SqlShowFunctions can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,882 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dql.SqlShowProcedures doesn't comprise a getter for subject databaseName
2024-08-05 19:06:21,882 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.dql.SqlShowProcedures doesn't comprise a setter for subject databaseName
2024-08-05 19:06:21,885 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.dql.SqlShowProcedures can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:21,890 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) class org.apache.flink.sql.parser.ddl.SqlReplaceTableAs doesn't comprise a setter for subject tableName
2024-08-05 19:06:21,893 INFO [org.apa.fli.api.jav.typ.TypeExtractor] (pool-10-thread-1) Class class org.apache.flink.sql.parser.ddl.SqlReplaceTableAs can't be used as a POJO kind as a result of not all fields are legitimate POJO fields, and should be processed as GenericType. Please learn the Flink documentation on "Data Types & Serialization" for particulars of the impact on efficiency and schema evolution.
2024-08-05 19:06:25,192 INFO [org.apa.fli.cli.pro.res.RestClusterClient] (Flink-RestClusterClient-IO-thread-1) Submitting job 'insert-into_ffbaaaacaaaaaaaaafaaaabafafafbaebfaa.yyya.abc' (5854f94a78a2ef2e79b264b97db11592).
2024-08-05 19:06:41,151 INFO [org.apa.fli.cli.pro.res.RestClusterClient] (Flink-RestClusterClient-IO-thread-4) Efficiently submitted job 'insert-into_ffbaaaacaaaaaaaaafaaaabafafafbaebfaa.yyya.abc' (5854f94a78a2ef2e79b264b97db11592) to 'http://127.0.0.1:8081'.
Information Verification
To confirm the success of the ingestion and supply to the configured goal databases, use the next MongoDB instructions.
Anticipated Consequence
You must see two information added to a set known as “data
” in a database known as “yyya
” equivalent to configured worth configuration.config.database.
Anticipated Output
yyya> db.knowledge.discover()
[
{
_id: ObjectId("66b924b40c964128eb12400a"),
id: 1,
name: 'name_1',
age: 46,
addr: { email: 'name_1@email.com', phone: '123' }
},
{
_id: ObjectId("66b924b40c964128eb12400b"),
id: '2',
name: 'name_2',
age: 55,
addr: { email: 'name_2@email.com', phone: '1234' }
}
]
yyya>
Conclusion
In conclusion, the trendy knowledge pipeline structure offers high-scale knowledge ingestion, knowledge elasticity, ordering, and knowledge consistency. An information pipeline structure can modernize a company’s knowledge wants for driving insights, analytics, ETL/ELT use instances, and machine studying infrastructure, amongst different use instances on this digital age.