Nested data in Parquet with Python
I believe this feature has finally been added in arrow/pyarrow 2.0.0:
https://issues.apache.org/jira/browse/ARROW-1644
https://arrow.apache.org/docs/python/json.html
Implementing the conversions on both the read and write path for arbitrary Parquet nested data is quite complicated to get right -- implementing the shredding and reassembly algorithm with associated conversions to some Python data structures. We have this on the roadmap in Arrow / parquet-cpp (see https://github.com/apache/parquet-cpp/tree/master/src/parquet/arrow), but it has not been completed yet (only support for simple structs and lists/arrays are supported now). It is important to have this functionality because other systems that use Parquet, like Impala, Hive, Presto, Drill, and Spark, have native support for nested types in their SQL dialects, so we need to be able to read and write these structures faithfully from Python.
This can be analogously implemented in fastparquet as well, but it's going to be a lot of work (and test cases to write) no matter how you slice it.
I will likely take on the work (in parquet-cpp) personally later this year if no one beats me to it, but I would love to have some help.