PyArrow provides its own filesystem abstraction optimized for Arrow/Parquet workflows with zero-copy integration.
See @data-engineering-storage-authentication for S3, GCS, Azure credential configuration.
Native Arrow filesystem integration with PyArrow. Optimized for Parquet workflows, zero-copy data transfer, predicate pushdown, and column pruning. Covers S3, GCS, HDFS with PyArrow datasets. Source: legout/data-platform-agent-skills.
Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/legout/data-platform-agent-skills --skill data-engineering-storage-remote-access-libraries-pyarrow-fs Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw