Reverse Dependencies of hdfs
The following projects have a declared dependency on hdfs:
- ai-flow-nightly — An open source framework that bridges big data and AI.
- alo7-airflow — Programmatically author, schedule and monitor data pipelines
- apache-airflow — Programmatically author, schedule and monitor data pipelines
- apache-airflow-providers-apache-hdfs — Provider package apache-airflow-providers-apache-hdfs for Apache Airflow
- apache-airflow-zack — Programmatically author, schedule and monitor data pipelines
- apache-beam — Apache Beam SDK for Python
- apache-beam-ai2 — A FORK! for testing with different dill version
- atelierflow — An ML pipeline using apache beam for run experiments
- cacheless-airflow — Programmatically author, schedule and monitor data pipelines
- custom-workflow-solutions — Programmatically author, schedule and monitor data pipelines
- dbnd-hdfs — Machine Learning Orchestration
- dvc-webhdfs — webhdfs plugin for dvc
- edu-airflow — Programmatically author, schedule and monitor data pipelines
- etl-ml — etl_ml is a tools could etl origin excel or csv dirty data and send data to ftp or server and insert data to hive database and load data from jump hive make feature project machine learning model train and jump the jump machine to connect hive get hive data to pandas dataframe
- file-io-utilities — Utils for file IO operations in Alida.
- functionalizer — A PySpark implementation of the Blue Brain Project Functionalizer
- GaiaXPy — Utilities to handle BP/RP (XP) Gaia low-resolution spectra as delivered via the archive
- geniusrise-databases — listeners bolts for geniusrise
- goe-framework — no summary
- hdfs-docling-analyze — A library for analyzing files from HDFS and saving results to MongoDB
- hdfs-kerberos-no-verify — A custom of from hdfs.ext.kerberos import KerberosClient from hdfs package that allow you to use with self-signed https url.
- importable — Allows to import zip-compressed Python package by URL (http, hdfs).
- kcevan — evan's private utils.
- kedro-datasets — Kedro-Datasets is where you can find all of Kedro's data connectors.
- khalinox — no summary
- lvfs — Convenient high level file IO across multiple protocols
- megfile — Megvii file operation library
- metatreedb — Metatree is a DBMS that uses the filesystem itself as a tree-structured database.
- mlflow-webhdfs — MLflow WebHDFS Plugins
- nalaf — Natural Language Framework, for NER and RE
- onetl — One ETL tool to rule them all
- oracle-ads — Oracle Accelerated Data Science SDK
- pano-airflow — Programmatically author, schedule and monitor data pipelines
- pylogrotate — Logrotate in Python
- pysparkling — Pure Python implementation of the Spark RDD interface.
- robotslacker-sqlcli — SQL Command test tool, use JDBC/ODBC
- robotslacker-sqlcli-noodbc — SQL Command test tool, use JDBC
- robotslacker-testcli — Test Command tool
- sparklightautoml — Spark-based distribution version of fast and customizable framework for automatic ML model creation (AutoML)
- sparklightautoml-dev — Spark-based distribution version of fast and customizable framework for automatic ML model creation (AutoML)
- ssh-jump-hive — ssh_jump_hive is a tools could jump the jump machine to connect hive get hive data to pandas dataframe
- target-jsonl-webhdfs — Singer.io target for writing JSON Line files via webhdfs
- TDY-PKG — its an implimentation of TF-2 , Detectron and yolov5
- TDY-PKG-saquibquddus — its an implimentation of TF-2 , Detectron and yolov5
- TFOD-Automatic — Automated Object detection for Beginner using python and Tensorflow
- tikit-en — Kit for TI PLATFORM
- tikit-test — Kit for TI PLATFORM
- wiseprophet — WiseProphet Platform package creation written by wiseitech
- wp-library — WiseProphet Platform package creation written by wiseitech
- xedro — Kedro helps you build production-ready data and analytics pipelines
1