You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Aug 13, 2020. It is now read-only.
That would allow the user to load very large parquet files but pay only for what is actually used (similarly to what vroom does https://github.com/r-lib/vroom)
what do you think?
Thanks!
The text was updated successfully, but these errors were encountered:
great idea!! maybe you should work with Jim Hester (@jimhester, vroom author) to get a single package that handles csv + parquet super fast? that would be a killer package in my opinion! and more dev are needed to fix bugs and other inefficiencies. what do you think?
Check out the altrep branch in this repo... for now, it materialises everything at once, but things like this should no longer read any unrelated payload data:
Hi,
Assuming that is (even technically) possible, it would be useful to have the data indexed (but not loaded yet in the RAM) like in
sparklyr
(see https://www.rdocumentation.org/packages/sparklyr/versions/1.0.2/topics/spark_read_parquet)That would allow the user to load very large parquet files but pay only for what is actually used (similarly to what
vroom
does https://github.com/r-lib/vroom)what do you think?
Thanks!
The text was updated successfully, but these errors were encountered: