Expand my Community achievements bar.

Join us for the next Community Q&A Coffee Break on Tuesday April 23, 2024 with Eric Matisoff, Principal Evangelist, Analytics & Data Science, who will join us to discuss all the big news and announcements from Summit 2024!
SOLVED

Data Warehouse - requests gzip compression

Avatar

Level 3

Hi,

 

we are using data warehouse requests and everything works fine. 

Unfortunately we put a lot of ressources to unzip the exports

 

Is there plan to change the compression type or add gzip compression in the next future?

Like in Data Feeds, where bigger files are automatically splitted into smaller ones?

 

ZIP it's not a supported format by Snowflake, Athena, etc .... 

Instead of unzip -> gzip -> load, we would like to load the exports directly.

 

Regards,

Piotr

 

1 Accepted Solution

Avatar

Correct answer by
Level 10

Hi Piotr,

 

Out of curiosity, what is the average file size (.zip) and export frequency?

 

P.S.: I would advise to submit it as an idea to get votes for this feature.

View solution in original post

5 Replies

Avatar

Correct answer by
Level 10

Hi Piotr,

 

Out of curiosity, what is the average file size (.zip) and export frequency?

 

P.S.: I would advise to submit it as an idea to get votes for this feature.

Avatar

Level 3

Hi Adrey,

 

the average size about 500MB, but 30% of the exports are beyond 1GB -> up to 1.5GB in a single file for each of our reporting suites.

The exports are scheduled on hourly basis. Currently we divide those with segments into smaller ones, but still it's a lot of overhead:

- to manage those exports (like add/remove columns)

- unzip and gzip 

 

Ok, I'll submit this as an idea.

 

Thx!  

 

 

 

Avatar

Level 10
500+ Mb on hourly basis — is it referring to Data Warehouse or Data Feeds?

Avatar

Level 3
500+ MB for Data Warehouse, for all reporting suites it's about 2-3 GB hourly