
Ditutup
Disiarkan
I want a complete, production-style data engineering project built on Azure and Databricks that I can showcase as a real-world reference preferably in healthcare or retail domain. The solution should ingest at-scale public data sets (think TPC transaction benchmarks for retail, Synthea synthetic EHR data for healthcare, or a clever blend of both) and drive two to three meaningful business use cases from raw landing all the way to presentation. I am purposely leaving the exact scenarios open so you can propose what will demonstrate the most insight, but they must be substantial enough to feel like something a modern data team would support in production. Data quality is my top priority. I need lineage, validation rules, automated tests, and observable metrics baked in from day one—Great Expectations, Delta Live Tables expectations, or comparable frameworks are welcome, as long as quality gates are visible in the monitoring layer. Scope to cover: • Architecture design diagram with clear component rationale (Azure Data Lake, Databricks, Delta, Unity Catalog, etc.). • Reproducible code (Python / PySpark, notebooks or repos) with CI/CD instructions. • Ingestion pipelines (batch or streaming), curated layers, and serving tier (SQL endpoints, Power BI, or dashboards of your choice). • Integrated monitoring, alerting, and cost-aware observability using native Azure tools or open-source add-ons. • End-to-end test suite: unit, integration, and data quality tests triggered via pipelines. • Comprehensive markdown that walks through setup, architecture, and business logic. Acceptance criteria: 1. Pipelines run on my Azure subscription with a simple deploy script. 2. Data quality reports surface failed expectations in a dashboard or log analytics workspace. 3. Each business use case produces a consumable output (table, visualization, or API) that confirms value.
ID Projek: 40333496
19 cadangan
Projek jarak jauh
Aktif 12 hari yang lalu
Tetapkan bajet dan garis masa anda
Dapatkan bayaran untuk kerja anda
Tuliskan cadangan anda
Ianya percuma untuk mendaftar dan membida pekerjaan
19 pekerja bebas membida secara purata ₹1,066 INR/jam untuk pekerjaan ini

Your data quality requirement will fail if you treat Great Expectations as an afterthought bolted onto finished pipelines. Most teams discover validation gaps only after bad data corrupts downstream dashboards, forcing expensive rollbacks. I've seen this pattern destroy three healthcare analytics projects where lineage looked perfect on paper but missing null checks in the bronze layer cascaded into compliance violations. Before architecting the medallion flow, I need clarity on two constraints. First, what's your Azure spend ceiling for the POC - are we optimizing for serverless Databricks SQL or can we provision dedicated clusters for streaming workloads? Second, does your use case require HIPAA-aligned audit trails, or is this purely a technical showcase without PHI sensitivity requirements? The answer changes whether we use Unity Catalog's fine-grained access controls or simpler RBAC patterns. Here's the architectural approach: - AZURE DATA LAKE + DELTA: Implement zone-based storage (raw/curated/presentation) with Delta time travel enabled so you can audit every transformation and roll back corrupt batches without losing history. - DATABRICKS + DELTA LIVE TABLES: Build declarative pipelines with inline expectations that auto-quarantine rows failing business rules, surfacing quality metrics in real-time dashboards rather than buried log files. - PYSPARK + GREAT EXPECTATIONS: Write parameterized test suites that validate schema drift, referential integrity, and domain constraints before promoting data between layers, triggered via Azure DevOps pipelines on every merge. - UNITY CATALOG + LINEAGE: Tag every table with business glossary terms and capture column-level lineage so analysts can trace a Power BI metric back to the source API without reverse-engineering notebooks. - CI/CD + TERRAFORM: Package the entire stack as infrastructure-as-code with automated deployment scripts that spin up workspaces, mount storage, and configure monitoring in under 20 minutes on a fresh subscription. I've built similar reference architectures for two healthcare SaaS platforms that passed SOC2 audits on first attempt because quality gates were non-negotiable from the bronze layer forward. Let's schedule a 20-minute call to align on cost guardrails and compliance scope before I propose the specific retail/healthcare use cases.
₹900 INR dalam 30 hari
4.4
4.4

I will build your end-to-end Azure + Databricks data engineering project with production-grade pipelines, Delta architecture, and strong data quality (Great Expectations/DLT) integrated with monitoring and CI/CD. You’ll get a fully deployable, well-documented solution with real-world datasets, actionable business use cases, and dashboards that clearly demonstrate data value and reliability.
₹1,000 INR dalam 40 hari
3.1
3.1

I’ll build you a production-grade Azure + Databricks data engineering project that actually looks like something a real data team would run—not a toy demo. I’ll design a scalable architecture (ADLS, Delta Lake, Unity Catalog) with robust ingestion pipelines using PySpark, handling large public datasets like Synthea or retail benchmarks. Data quality won’t be an afterthought—I’ll integrate Great Expectations/Delta Live Tables with automated tests, lineage, and observable metrics wired into monitoring dashboards. You’ll get CI/CD-ready code, deploy scripts, and clear documentation. I’ll also define 2–3 strong business use cases with dashboards (Power BI/SQL endpoints) that prove real value. Everything will be reproducible, testable, and cleanly handed over.
₹1,500 INR dalam 40 hari
2.8
2.8

Hi, I’m Karthik, a Data & Cloud Architect with 15+ years of experience delivering production-grade Azure + Databricks data platforms. I can build a complete, portfolio-ready end-to-end pipeline in healthcare (Synthea EHR) or retail (TPC-style), covering real-world use cases like patient risk scoring, cohort analysis, or customer segmentation & sales forecasting. **Proposed Solution:** • Architecture: Azure Data Lake Gen2 + Databricks (Delta Lake, Unity Catalog) + Azure Data Factory / Workflows • Medallion design (Bronze/Silver/Gold) with scalable batch/stream ingestion • Data quality-first approach using Delta Live Tables expectations + Great Expectations • Full lineage, validation rules, and observable metrics via Azure Monitor & Log Analytics • CI/CD with GitHub Actions / Azure DevOps (reproducible deployment scripts) • Serving layer via Databricks SQL + Power BI dashboards **Deliverables:** • Architecture diagram with rationale • PySpark notebooks/repos with modular pipelines • Automated tests (unit, integration, data quality gates) • Monitoring, alerting & cost observability • Detailed markdown documentation (setup → business insights) Each use case will output actionable insights (dashboards/tables/APIs) proving business value. I’ve implemented similar enterprise data platforms with strong governance, quality, and scalability. Happy to propose the best domain & use cases aligned to your goals. Warm Regards, Karthik Resonite Technologies
₹1,330 INR dalam 40 hari
4.1
4.1

Hello, I can build a complete, production-grade Azure + Databricks data engineering project suitable for a real-world portfolio, using healthcare, retail, or a hybrid dataset (e.g., Synthea + TPC). The solution will include a full medallion architecture (Bronze/Silver/Gold) using Azure Data Lake, Databricks, Delta Lake, and Unity Catalog with proper governance and lineage. It will cover end-to-end pipelines for ingestion (batch/streaming), transformation, and serving layers, along with 2–3 meaningful business use cases that produce dashboards, tables, or APIs. Data quality will be a core focus, with built-in validation using Delta Live Tables or Great Expectations, plus visible failure tracking and quality metrics. I will also include CI/CD deployment, PySpark notebooks, architecture diagram, testing framework (unit + integration + data tests), and Azure monitoring (Log Analytics/Azure Monitor). The entire project will be deployable on your Azure subscription with a simple setup script and clear documentation. Regards, Nainsi
₹1,100 INR dalam 20 hari
0.0
0.0

Hi, I’m a data engineer experienced with Python, PySpark, Azure, and Databricks, and I can build a production-ready data platform for healthcare or retail datasets. My solution will include end-to-end pipelines, data quality checks with Great Expectations/Delta Live Tables, and actionable insights through dashboards or APIs. I’ll deliver deployable, well-documented code with reproducible pipelines and business use cases that demonstrate real-world impact.
₹1,000 INR dalam 40 hari
0.0
0.0

I am a data entry expert who ensures flawless results. If any errors occur, I rework the task multiple times to guarantee perfection." Would you like me to refine this into a formal bio for a professional profile?
₹850 INR dalam 40 hari
0.0
0.0

I am highly impressed by your focus on data quality and observability; it is exactly how modern enterprise-grade systems should be built. I strongly agree that a real-world reference must go beyond simple ETL and include robust governance and automated testing. I am prepared to build a "Medallion Architecture" on Azure using Databricks and Unity Catalog that serves as a gold standard for production-ready data engineering. For this project, I propose a Healthcare Analytics Hub using Synthea EHR data. We will drive two critical business cases: Patient Readmission Risk Profiling and Operational Efficiency Analytics. The solution will feature a "Quality-First" ingestion layer using Delta Live Tables (DLT) Expectations for row-level validation and Great Expectations for automated post-load profiling. This ensures that every byte of data is validated against strict quality gates before reaching the Gold layer. To ensure transparency and reproducibility, I will deliver a fully automated CI/CD pipeline (GitHub Actions or Azure DevOps) and a Power BI Observability Dashboard that surfaces data lineage and health metrics directly from your metadata. You will receive a clean, modular repository with a "one-click" deployment script, allowing the entire environment to be spun up on your Azure subscription seamlessly. Hourly Rate: ₹1,500.00 Weekly Limit: 40 Hours
₹1,500 INR dalam 40 hari
0.0
0.0

Hello, I understand you're looking for a complete data engineering project on Azure and Databricks that emphasizes high data quality and tangible business use cases in healthcare or retail. My expertise in Python, Databricks, and Azure services positions me well to meet your requirements. For the architecture, I would propose leveraging Azure Data Lake for data storage, Databricks for processing, and Delta Live Tables for managing data pipelines. We’ll design ingestion pipelines capable of handling both batch and streaming data from public datasets. Implementing comprehensive data quality measures will be a priority; we’ll use frameworks like Great Expectations to ensure lineage, validation, and automated testing are integral to the design. The workflow will include CI/CD practices for reproducible code in Python/PySpark, integrated monitoring via Azure tools, and an end-to-end test suite to validate every asset before deployment. The final deliverables will contain a clear architecture diagram, detailed markdown documentation, and outputs that demonstrate value through meaningful visualizations or APIs. Final details can be discussed further, ensuring all aspects align with your vision. Best regards, Satya
₹1,000 INR dalam 7 hari
0.0
0.0

I have all the required skills and expertise of azure databricks and microsoft azure. As, I am currently working on the same tech stacks. I have build an end-to-end data pipeline for a medical care project coming from SAP system and performs the data handling all through unity catalog enabled databricks workspace by creating external delta tables. Also deployed it through Databricks Assests Bundle in Azure Devops(ADO). Also applied the alert notifications and auditing properly to have a constant eye on the pipeline
₹1,000 INR dalam 20 hari
0.0
0.0

Hello, I’m excited to help you create a professional, ATS-friendly resume that stands out and increases your chances of getting selected. I specialize in designing resumes that are not just visually clean but also optimized for recruiters and hiring systems (ATS). I focus on highlighting your strengths, skills, and achievements in a way that grabs attention within seconds. ? What you’ll get: ✔ Clean and modern resume design ✔ ATS-friendly formatting (highly important) ✔ Strong, impactful content writing ✔ Proper structure for maximum readability ✔ Quick delivery with revisions I recently created resumes for clients that are tailored for jobs, internships, and freelancing platforms, ensuring they look professional and competitive. I understand what recruiters look for, and I’ll make sure your resume reflects your true potential. I’m ready to start immediately and deliver high-quality work. Let’s create a resume that gets you hired ? Best regards, Shivraj Maske
₹800 INR dalam 30 hari
0.0
0.0

As a seasoned data engineer with over four years of experience, I have honed my skills in delivering end-to-end data pipeline solutions, precisely like what you're looking for. My experience ranges from building data warehouses to designing robust ETL systems and automating data flows using PySpark and SQL. Notably, I have created fastapis that manage, trigger, and update Spark pipelines in Kubernetes; this provides me with excellent insights into deploying the solutions you seek on Azure. Your project's focus on monitoring is something I take very seriously; thus I'm proficient in incorporating cost-aware observability measures into my solutions using native Azure tools or open-source add-ons. This enables me to provide failsafe alerts through comprehensive dashboards and log analysis workspaces where any failed expectations can be pinpointed. Importantly, I am known for ensuring that all my deliverables provide valuable outputs - be it tables, visualizations or APIs - that confirm the value of the project like you require.
₹925 INR dalam 28 hari
0.0
0.0

Hi, I will build a production-style Azure Databricks data engineering solution in your chosen domain (healthcare with Synthea or retail with TPC-style transactions), delivering 2–3 end-to-end business use cases from raw landing to dashboards. This will include a clear architecture (ADLS, Databricks, Delta, Unity Catalog), reproducible PySpark notebooks with CI/CD guidance, bronze–silver–gold layers, and a serving tier via SQL endpoints and Power BI (or similar). I will implement robust data quality with expectations, lineage, automated tests, and Azure-native monitoring and alerting, plus markdown documentation so you can easily run and showcase the project in your own subscription. Estimated timeline is 10–14 days , depending on final domain and use case depth. Which domain and top 2–3 business outcomes do you most want this portfolio project to highlight for hiring managers or stakeholders? If this matches your vision, I’d be glad to refine the scenarios with you and then deliver a clean, observable solution that truly looks and behaves like modern production data engineering. Thanks
₹2,000 INR dalam 40 hari
0.0
0.0

I am working as a Azure data engineer for j&j. If you want to contact me. This is my number 6304572145
₹1,000 INR dalam 40 hari
0.0
0.0

Hi, I am a SQL Developer & Data Engineer with 4+ years hands-on experience building ETL pipelines in healthcare domain (large US Digital Healthcare System with 28+ modules for Stanford Hospitals initiative). I will deliver a complete Azure + Databricks end-to-end production-style pipeline using Synthea synthetic EHR data. What I will provide: Full Medallion architecture (Bronze → Silver → Gold) on Azure Data Lake, Databricks, Delta Lake, Unity Catalog PySpark ingestion, transformation, and quality checks 2-3 business use cases with dashboards Lineage, data quality, monitoring, CI/CD, tests, diagrams, and complete documentation I can start immediately and make it fully reproducible on your Azure subscription. Happy to share a quick plan. Best regards, Mallela Shaheena SQL Developer | Data Engineer | Azure Data Factory | PySpark
₹850 INR dalam 40 hari
0.0
0.0

HI there, As a seasoned DevOps Engineer with deep experience in cloud technologies, Azure in particular, I'm well-prepared to take on the challenge of designing and delivering your Azure and Databricks End-to-End Pipeline. Having worked on various data engineering projects, I understand the critical role that data quality plays in the success of any solution. Leveraging my experience with CI/CD and cloud platforms, I can ensure that your project incorporates strong lineage, validation rules, automated tests, and observable metrics from day one. Moreover, my skill-set extends beyond just the technical aspects; I bring with me a unique background in biotechnology which gives our partnership an edge for this project in healthcare or retail domains. This marries my expertise in cloud infrastructure and automation with my deep understanding of impactful use cases. Together, we can identify the most meaningful business scenarios that align with modern data team's needs and demonstrate real-world applicability for your project. Lastly, my proficiency as a content writer will augment your deliverables. I'll readily create comprehensive markdowns elucidating the setup, architecture, and business logic; involving your stackholders throughout the process while making sure that a simple deploy script could get things up and running on your Azure subscription. Let's get started Best Regardas Laiba
₹750 INR dalam 40 hari
0.0
0.0

1 Designed and developed scalable ETL pipelines using Azure Databricks, PySpark, and ADF 2 Built Medallion Architecture (Bronze, Silver, Gold layers) 3 Reduced data processing time by 40% 4 Optimized Spark jobs for better performance 5 Developed Power BI dashboards and KPIs 6 Implemented Row-Level Security (RLS) 7 Created DAX measures and analytics reports 8 Ensured data quality and reliability
₹1,000 INR dalam 40 hari
0.0
0.0

Hi, I read your project regarding Azure and Databricks End-to-End Pipeline and I can definitely help with that. I am a with 4.5+ years of experience specializing in Azure Databricks and Azure services , Implementing Pipeline. I can resolve this by pypsark code and databricks functionality, create batch process and schedule as required timestamp . When do you need this finished? I can start immediately. Required in details to project start.
₹1,000 INR dalam 40 hari
0.0
0.0

Hi, I’m a Data Engineer with full-time industry experience in Azure, Databricks, Python, and SQL. I’ve worked on end-to-end data pipelines, from ingestion to reporting. I’m new to freelancing, so I’m offering a lower-than-standard quote to build my profile and gain good reviews. But I’m fully committed to delivering a high-quality, production-ready solution with proper architecture, data quality checks, and clear documentation. I’ll make sure the project is scalable, easy to deploy, and delivers real business value. Looking forward to working with you.
₹750 INR dalam 40 hari
0.0
0.0

new delhi, India
Ahli sejak Nov 27, 2017
$15-25 USD / jam
$15-25 CAD / jam
₹750-1250 INR / jam
₹12500-37500 INR
₹600-1500 INR
$750-1500 USD
$30-250 USD
$10000-20000 USD
$15-25 USD / jam
$15-25 USD / jam
₹37500-75000 INR
$250-750 USD
₹100-400 INR / jam
₹400-750 INR / jam
$15-25 USD / jam
₹600-601 INR
$10-30 USD
€8-20 EUR
$250-750 USD
₹1500-12500 INR
$250-750 USD
$30-250 USD
₹750-1250 INR / jam
$5000-10000 USD
₹750-1250 INR / jam