-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
need help to Store fluentd logs in azure datalake as parquet.gz format #19
Comments
@oleewere , I will appreciate if you can provide some help here ,Thanks |
@wasifshareef - s3 has that implementation, here it does not but i will try to add that feature on the weekend |
@oleewere , Thanks very much. I appreciate your help and support here |
you can check this change: #20 - maybe test the file on your code if you think |
Hi @oleewere , I was able to update below location with azurestorage_gen2_compressor_parquet.rb file you shared. Can you please share sample conf file to know how you updated the store as value there in conf file. root@3982bba4b341:/opt/td-agent/lib/ruby/gems/2.7.0/gems/fluent-plugin-azurestorage-gen2-0.3.5/lib/fluent/plugin# ls -lrt |
Does this look good to you. Do we need to define schema type and file aswell ? . i tried with this conf parameters ,but infortunately its still save files in azure datalake as .txt . I will appreciate if you can provide your inputs here , Thanks!
|
did you see any of these logs? "Use 'text' instead" ? (as only textcompressor should use that extension) |
I have tried but its still sending logs to azure as normal txt files . Sure , that will be really helpful if you can test and update here |
Hi @oleewere , I will appreciate if you can help to fix the parquet format store issue |
hi @wasifshareef - in the next few days i will have more time so i think i can take a look |
@oleewere , Thanks very much, I appreciate your support here ! |
@wasifshareef - moved changes to out_azurestorage_gen2.rb on the PR, can you check that implementation again ? also not it requires columnify to be installed (based on the s3 implementation) |
@oleewere , |
@wasifshareef - updated the PR to provided an example with simple json (similar configs as for the s3 plugin) |
HI @oleewere , Thanks for providing an example . 2024-07-07 15:29:50 +0000 [warn]: #0 bad chunk is moved to /tmp/fluent/backup/worker0/object_8ac/61ca9f6fde3786d894f458f6fe3b0b49.log |
@wasifshareef with the sample (+ installed columnify) it's looking good for me,
wondering if any version is different on your env that could matter ( ? ), seems the error coming from columnify when it called on the local tmp file |
Hi,
I will appreciate if you can help me to conform if fluentd-plugin-azurestorage-gen2 supports fluentd logs shipped to azure datalake as perquet.gz format.
I tried with below configuration for store but it doesnot store files in perquet.gz . I believe perquet.gz format is supported with fluentd-s3 plugin
@type azurestorage_gen2
time_as_integer true
##################################################################
azure_oauth_tenant_id xxxxxxxx
azure_oauth_app_id xxxxxxxxx
azure_oauth_secret xxxxxxxxx
azure_oauth_identity_authority https://login.microsoftonline.com
azure_storage_account xxxxx
azure_container xxxxx
auto_create_container true
store_as gzip
The text was updated successfully, but these errors were encountered: