Remote monitoring and predictive maintenance with Azure Databricks and CosmosDB
-
Updated
Dec 11, 2018
Remote monitoring and predictive maintenance with Azure Databricks and CosmosDB
Reading the Avro files created by Event hubs using Spark
Notebook with data ETL use cases with Spark
Connect to an Azure Databricks workspace without using a PAT within a Github pipeline
A midterm on breadth first search, map reduce, and PySpark transformations
An Apache Spark course based on Spark: The Definitive Guide
I have forked this template to implement end to end Machine Learning Life cycle on Databricks Lakehouse
Real estate sales predictions and analytics
Implementation of the "CCF: Fast and Scalable Connected Component Computation in MapReduce" paper with Spark. Study of its scalability on several datasets using various clusters' sizes on Databricks and Google Cloud Platform (GCP)
Data pipeline that processes Formula1 data with Azure Databricks, DeltaLake, and Azure Data Factory
The primary objective of this study is to explore the feasibility of using machine learning algorithms to classify health insurance plans based on their coverage for routine dental services. To achieve this, I used six different classification algorithms: LR, DT, RF, GBT, SVM, FM(Tech: PySpark, SQL, Databricks, Zeppelin books, Hadoop, Spark-Submit)
Ingestão de dados do Olist em formato CSV para as camadas Raw, Bronze, Silver e Gold
The data engineering team focuses on establishing a robust and reliable data pipeline. We use Kafka to manage the data streaming topics and later process and consume this data with the help of Spark.
Neste repositório trabalharemos com processamento de dados usando Spark.
Batch & streaming data pipelines built using Databricks with Pyspark and modeled the data into star schema to analyze in PowerBI, Formula-1 racing data from multiple data sources, APIs.
This Repo Contains Azure Data Engineering Projects
Add a description, image, and links to the databricks topic page so that developers can more easily learn about it.
To associate your repository with the databricks topic, visit your repo's landing page and select "manage topics."