Lists (1)
Sort Name ascending (A-Z)
Stars
PyScript is an open source platform for Python in the browser. Try PyScript: https://pyscript.com Examples: https://tinyurl.com/pyscript-examples Community: https://discord.gg/HxvBtukrg2
Change data capture for a variety of databases. Please log issues at https://issues.redhat.com/browse/DBZ.
Generation of diagrams like flowcharts or sequence diagrams from text in a similar manner as markdown
dbt enables data analysts and engineers to transform their data using the same practices that software engineers use to build applications.
The easiest way to get started with PostgreSQL on the Mac
Python Module for Tabular Datasets in XLS, CSV, JSON, YAML, &c.
💨 Fast, Async-ready, Openapi, type hints based framework for building APIs
A platform to build and run apps that are elastic, agile, and resilient. SDK, libraries, and hosted environments.
Redshift JDBC Driver. It supports JDBC 4.2 specification.
Create committing rules for projects 🚀 auto bump versions ⬆️ and auto changelog generation 📂
CLI tool that can execute SQL queries on CSV, LTSV, JSON, YAML and TBLN. Can output to various formats.
Apache Flink Kubernetes Operator
Utility to compare data between homogeneous or heterogeneous environments to ensure source and target tables match
The Lineage Analysis system for FlinkSQL supports advanced syntax such as Watermark, UDTF, CEP, Windowing TVFs, and CTAS.
Amazon Redshift Utils contains utilities, scripts and view which are useful in a Redshift environment
SeaTunnel is a next-generation super high-performance, distributed, massive data integration tool.
Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
Apache DolphinScheduler is the modern data orchestration platform. Agile to create high performance workflow with low-code
jgrapht / jgrapht
Forked from lingeringsocket/jgraphtMaster repository for the JGraphT project
AWS Glue Libraries are additions and enhancements to Spark for ETL operations.
Deequ is a library built on top of Apache Spark for defining "unit tests for data", which measure data quality in large datasets.