Crack Detection model using yolov7
-
Updated
Jul 2, 2023 - Jupyter Notebook
Crack Detection model using yolov7
A movie recommender written in Go that suggests movies considering various factors within a particular dataset, encompassing users, movies, and movie ratings.
BigQuery data pipeline with dbt, Spark, Docker, Airflow, Terraform, GCP
Setting up a Spark cluster in a Docker environment for improved repeatability and reliability. This project includes a simple transformation on a dataset containing approximately 31 million rows.
Solved tasks of the master's degree courses of speciality "Algorithms and Systems for Big Data Processing".
"Provides tools for parallel pipeline processing of large data structures
Software basati su metodi di intelligenza artificiale per l'automazione dell'analisi di big data.
Degree diploma project
A Docker Compose Template to deploy Airflow with sync from a remote repository
Tech blog / notes from my various endeavours and exploits
Building Data Lake and ETL pipelines using Amazon EMR, S3, and Apache Spark
Experiment to record as much data as possible in a given amount of time using a distributed timeseries database.
rock-solid pillars for enterprise-grade solutions
Data modeling with Cassandra, building Data Warehouse using Redshift and creation of Data Lake using Spark and Airflow
Analyzing classified ads data from the used motorcycles market. Tasks involve utilizing Redis Bitmaps for analytics on seller actions and MongoDB for analyzing bike listings. Includes data installation, cleaning, and analysis.
datasets-toolbox are some scripts usefull to generate, transfom and valid large dataset files, not openable with editor because too large. datasets-toolbox provide also a ping script.
Implementation of algorithms for big data using python, numpy, pandas.
Collection of homework (mostly Spark-based) from the course "Big Data Computing" - University of Padua.
Add a description, image, and links to the big-data-processing topic page so that developers can more easily learn about it.
To associate your repository with the big-data-processing topic, visit your repo's landing page and select "manage topics."