is there a way to figure out what the uncompressed file size is for a parquet file compressed in snappy? I have a lot of parquet files in a HDFS directory and I'm trying to figure out if there is a way to calculate the file size if that data was uncompressed.
            Asked
            
        
        
            Active
            
        
            Viewed 2,202 times
        
    1 Answers
0
            
            
        You can just try to uncompress the data and see how much space it consumes. See How to decompress the hadoop reduce output file end with snappy?
Maybe there's a more elegant way I'm not aware of
        Community
        
- 1
 - 1
 
        Lior Chaga
        
- 1,424
 - 2
 - 21
 - 35