Reverse Dependencies of pyspark
The following projects have a declared dependency on pyspark:
- highcharts-stock — High-end Time Series Data Visualization for the Python Ecosystem
- histogrammar — Composable histogram primitives for distributed data reduction
- historian-query — Query regularized time series from raw historian data on Spark
- hops-petastorm — Petastorm is a library enabling the use of Parquet storage from Tensorflow, Pytorch, and other Python-based ML training frameworks. This is a fork of Petastorm that is compatible with Hops installations
- hopsworks — Hopsworks Python SDK to interact with Hopsworks Platform, Feature Store, Model Registry and Model Serving
- hotpp-benchmark — Evaluate generative event sequence models on the long horizon prediction task.
- HTF_PythonProjectTest — Sample Python Project for creating a new Python Module
- hummingbird-ml — Convert trained traditional machine learning models into tensor computations
- husqvarna-getl — An elegant way to ETL'ing
- hyperleaup — Create and publish Tableau Hyper files from Apache Spark DataFrames and Spark SQL.
- hyperopt — Distributed Asynchronous Hyperparameter Optimization
- ibis-framework — The portable Python dataframe library
- iguanas — Rule generation, optimisation, filtering and scoring library
- iMapHub — Library created to map two Dataset
- in-dbt-spark — Release for LinkedIn's changes to dbt-spark.
- inference-schema — This package is intended to provide a uniform schema for common machine learning applications, as well as a set of decorators that can be used to aid in web based ML prediction applications.
- infra2conn — Infra 2.0 Connections
- intake-hive — Intake Hive DataSource Plugin.
- intake-spark — Apache Spark plugin for Intake
- internet — Python library for working with IP addresses and other internet related functionalities
- ipyvizzu — Build animated charts in Jupyter Notebook and similar environments with a simple Python syntax.
- irf — irf
- japanAirAnalytics — This software is being developed at the University of Aizu, Aizu-Wakamatsu, Fukushima, Japan
- jayspytools — no summary
- jeremydimond.pymlga — Python machine learning by genetic algorithm
- jgbpacking — PACKAGE MODELO CLASIFICACION WINE
- jgbpackinx — PACKAGE MODELO CLASIFICACION WINE
- jhdfs4py — Convenient HDFS access using the Java HDFS client
- jmetalpy — Python version of the jMetal framework
- jobsworthy — no summary
- johnsnowlabs — The John Snow Labs Library gives you access to all of John Snow Labs Enterprise And Open Source products in an easy and simple manner. Access 10000+ state-of-the-art NLP and OCR models for Finance, Legal and Medical domains. Easily scalable to Spark Cluster
- johnsnowlabs-by-kshitiz — The John Snow Labs Library gives you access to all of John Snow Labs Enterprise And Open Source products in an easy and simple manner. Access 10000+ state-of-the-art NLP and OCR models for Finance, Legal and Medical domains. Easily scalable to Spark Cluster
- johnsnowlabs-my-mehmet — The John Snow Labs Library gives you access to all of John Snow Labs Enterprise And Open Source products in an easy and simple manner. Access 10000+ state-of-the-art NLP and OCR models for Finance, Legal and Medical domains. Easily scalable to Spark Cluster
- johnsnowlabs-tmp — The John Snow Labs Library gives you access to all of John Snow Labs Enterprise And Open Source products in an easy and simple manner. Access 10000+ state-of-the-art NLP and OCR models for Finance, Legal and Medical domains. Easily scalable to Spark Cluster
- json2spark-mapper — Maps JSON schema types to Spark SQL types
- jsonSpark — This is a wrapper package for pyspark to process json files. It pythonifies the json pyspark object.
- jupysql — Better SQL in Jupyter
- jupyterlab-sparkmonitor — Spark Monitor Extension for Jupyter Lab
- kagglebh — Analyze kaggle dataset: climate-change-earth-surface-temperature
- katonic — A modern, enterprise-ready MLOps Python SDK
- kcommon — Sample Python Project for creating a new Python Module
- kedro-datasets — Kedro-Datasets is where you can find all of Kedro's data connectors.
- kedro-great — Kedro Great makes integrating Great Expectations with Kedro easy!
- kedro-pandera — A kedro plugin to use pandera in your kedro projects
- kedro-popmon — Kedro Popmon makes integrating Popmon with Kedro easy!
- kgdata — Library to process dumps of knowledge graphs (Wikipedia, DBpedia, Wikidata)
- kgf — KGF feature engineering package
- Kivi — A risk control modeling tool python package.
- kloppy-spark — Spark Tools to work with Kloppy
- koalas — Koalas: pandas API on Apache Spark
- kobai-sdk — A package that enables interaction with a Kobai tenant.
- koheesio — The steps-based Koheesio framework
- kozmoserver-mllib — Spark MLlib runtime for KozmoServer
- kumuniverse — Data team shared library for accessing services
- kurveclient — A client for the Kurve backend.
- kustopy — Query and Ingestion Client for Azure using Python
- l2-data-utils — no summary
- labelspark — Labelbox Connector for Databricks
- lake-loader — no summary
- lakehouse-engine — A configuration-driven Spark framework serving as the engine for several lakehouse algorithms and data flows.
- laktory — A DataOps framework for building a lakehouse
- lale — Library for Semi-Automated Data Science
- langchain_1111_Dev_cerebrum — Building applications with LLMs through composability
- langchain-by-johnsnowlabs — Building applications with LLMs through composability
- langchain-xfyun — 在LangChain中流畅地使用讯飞星火大模型
- langchaincoexpert — Building applications with LLMs through composability
- langchainmsai — Building applications with LLMs through composability
- langchainn — Building applications with LLMs through composability
- langplus — Building applications with LLMs through composability
- layer-client — The Layer Client
- learnstdio — Learnstdio for Python
- lexios — Sample Python Project for creating a new Python Module
- liga — no summary
- ligavision — no summary
- lingualeo-sqlmesh — no summary
- litenai — Python library for Liten AI Data platform
- llm-explorer — A Lakehouse LLM Explorer. Wrapper for spark, databricks and langchain processes
- localstack-s3-pyspark — A CLI to configure pyspark for use with s3 on localstack
- loinchpo — A library to transform loinc measurements into hpo terms.
- longalpha-utils — no summary
- luisy — Framework to build data pipelines
- luminex — On the Fly ETL application
- luntaiDs — Make Data Scientist life Easier Tool
- luq89-pyspark-app-luq89 — Sample app in PyPI
- mAdvisor — An automated AI/ML solution from Marlabs
- maggy — Efficient asynchronous optimization of expensive black-box functions on top of Apache Spark
- manifestare — Simple, fluent, and descriptive API to explicit DataFrames expectations.
- mapGlobaltoLocal — Library created to map two Dataset
- markdown_frames — Markdown tables parsing to pyspark / pandas DataFrames
- mars-gym — Framework Code for the RecSys 2020 entitled 'MARS-Gym: A Gym framework to model, train, and evaluate recommendationsystems for marketplaces'.
- MD2K-Cerebral-Cortex — Backend data analytics platform for MD2K software
- mdspy — Mnubo Data Science Library
- mega-spark — Some computing tools about feature tool and ml model.
- meliora — Credit risk validation and development tools
- memoria — Python library of hashing and caching
- mercury-ml — A library for managing Machine Learning workflows
- merlin-batch-predictor — Base PySpark application for running Merlin prediction batch job
- mindlab — Data science toolbox
- Mits-AI-Dreamers — This is Hamming distance, Jaccard coefficient, overlap
- mk-feature-store — Python SDK for Feast