Direct Pipeline with No Transforms #4872
Unanswered
jermjensen
asked this question in
Q&A
Replies: 2 comments 1 reply
-
Try changing the "row set size" parameter (default 10.000) in your local pipeline run configuration. |
Beta Was this translation helpful? Give feedback.
1 reply
-
have you tried providing limits in your SQL query You could make this more dynamic through parameters/variables, something like: This last option will require you to build a loop to go over all the available 100 row result sets, but would give you more flexibility. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm curious to know if Hop's behavior is different between two types of pipelines: a pipeline with multiple transforms vs. a pipeline with no transforms - just a source and target. I'm running version 2.9.
I've been trying to meter the flow of 10,000 rows of data from a generic JDBC source piped into a SQL Server target. I've set a low fetch size parameter on the source connection string (100 rows), a pipeline row set size = 1000, and a commit of 1000 rows in my target.
I was hoping to batches of 100 rows flow from the source, buffer up to 1000 on machine running Hop, the push and commit those 1000 rows to SQL Server.
What (I think) I'm seeing is that all 10,000 rows are read from the source (no rows are buffered) and then sent to the target with a quick succession of commits every 1000 rows. I'm also seeing the pipeline taking a long time to go from "idle" to "running".
Do I need to look more into my JDBC documentation (maybe I'm not pulling back only 100 rows at a time) or does Hop not use the buffer when the pipeline is source-straight-to-target?
Beta Was this translation helpful? Give feedback.
All reactions