
Kapalı
İlan edilme:
Teslimde ödenir
I have an existing SAS program that handles end-to-end data processing for a single SQL Database source. The code cleans raw tables, applies a series of transformations, then produces several aggregated outputs that feed downstream reports. I now need the entire workflow re-implemented in PySpark running on Azure Databricks so I can retire the SAS environment and take advantage of Databricks’ scalability. You will receive: • The original .sas files with inline comments that explain each step • A data-dictionary of the SQL tables involved • Sample input/output datasets to verify parity What I’m expecting from you: 1. A well-structured Databricks notebook (or .py files) that reproduces the SAS logic for data cleaning, transformation, and aggregation. 2. Connection code that pulls directly from the same SQL Database and writes the final tables back, using best practices for credentials and Spark JDBC. 3. Clear in-line comments so my team can maintain the pipeline going forward. 4. A short hand-off session where we review the notebook, run through the full load, and confirm the outputs match the SAS results record-for-record. Acceptance criteria – All row counts and aggregation values match the SAS outputs (±0 where rounding is not involved). – Runtime in Databricks is comparable or faster than the current SAS batch. – Code follows PySpark and Databricks style conventions and can be scheduled as a job without manual intervention. If you have deep experience translating SAS data pipelines into PySpark and are comfortable working inside Azure Databricks, I’d love to work with you on this migration.
Proje No: 40043251
41 teklifler
Uzaktan proje
Son aktiviteden bu yana geçen zaman 1 ay önce
Bütçenizi ve zaman çerçevenizi belirleyin
Çalışmanız için ödeme alın
Teklifinizin ana hatlarını belirleyin
Kaydolmak ve işlere teklif vermek ücretsizdir
41 freelancer bu proje için ortalama $474 USD teklif veriyor

I, Sandeep, as the CEO and Founder of Digital Screencast, possess over 7 years of experience as a Full Stack Developer with a specialty in Data Processing and SQL. My expertise lies extensively in converting SAS data pipelines to other data processing technologies that most times are faster and more flexible than SAS, like PySpark (which I have deep experience with). Added to this, I am skilled in working with Azure Databricks that your project mandates. I have previously worked on similar projects with companies like Metlife GOSC, DXC Technologies, etc., where I successfully migrated their pipelines for better performance and scalability - exactly what you seek! Perusing through your project description, the deliverables: creating a well-structured Databricks notebook; establishing robust connections directly from your SQL Database; generating comprehensible and maintainable code; and conducting a smooth hand-off process are all areas within my forte. Furthermore, I will ensure that our work together goes beyond the proposed timeframe by providing an incessant four-day post-completion support period for any issues that may arise.
$300 USD 5 gün içinde
8,3
8,3

Hello, I can re-implement your entire SAS workflow in PySpark on Azure Databricks, delivering a clean, well-commented notebook or .py files that connect to your SQL Database, reproduce all data cleaning, transformations, and aggregations, and write final tables back securely with Spark JDBC. 1. Do you want the notebook structured as one full workflow or modular scripts per transformation stage? 2. Are there any existing Databricks clusters or should I recommend a cluster setup optimized for your dataset size? 3. Should we include automated validation comparing PySpark outputs to SAS results as part of the pipeline? Our experienced team specializes in migrating SAS pipelines to PySpark, ensuring record-for-record parity, optimized runtime, clear documentation, and production ready scheduling. We can also conduct a short hand-off session to review the full workflow and confirm output accuracy. The current bid amount is a placeholder to submit. Regards, Yasir LEADconcept PS: I can provide examples of prior SAS to PySpark migrations and Databricks pipelines on request.
$500 USD 7 gün içinde
7,8
7,8

Hi I specialize in migrating complex SAS data pipelines to PySpark on Azure Databricks while preserving exact business logic and output parity. The main technical risk in SAS-to-PySpark migrations is subtle behavioral differences in joins, aggregations, and numeric handling, and I address this by translating each step deterministically and validating results record-for-record. I have strong experience re-implementing SAS data cleansing and transformation logic using PySpark DataFrames and Spark SQL following Databricks best practices. I connect securely to SQL databases via Spark JDBC with proper credential management and optimized read/write strategies. I structure notebooks and modules to be readable, testable, and ready for job scheduling without manual steps. I also include detailed inline comments so your team can maintain and extend the pipeline confidently. The outcome is a scalable Databricks workflow that fully replaces SAS with equal or better performance. Thanks, Hercules
$500 USD 7 gün içinde
6,2
6,2

Hello Gopal R., I specialize in re-engineering SAS pipelines into PySpark on Azure Databricks, preserving logic 1:1 while achieving major runtime improvements. I can show you demo PySpark notebooks + SAS-to-Spark migration samples before we finalize. ✅ What I Will Deliver 1️⃣ End-to-end Databricks Notebook / .py pipeline • Full replication of your SAS cleaning, joins, filters, transforms • Aggregation logic reproduced with exact parity (record-for-record) • Modular, readable, version-controlled structure 2️⃣ SQL Database Integration (Best Practices) • Secure connection via Key Vault / scoped secrets • Optimized Spark JDBC reads/writes • Partitioning + caching for performance 3️⃣ Production-Ready ETL • Job-oriented notebook structure • Idempotent Spark code • Config-driven transformations 4️⃣ Documentation + Handoff • Clear inline comments • Comparison report: SAS vs PySpark outputs • Walkthrough call + job scheduling guidance ? Techniques Used ✔ PySpark DataFrame API ✔ Window functions, UDFs, broadcast joins ✔ Databricks Workflows ✔ SQL DB connectivity + Delta Lake ✔ Unit testing with sample SAS outputs ? Relevant Projects SAS2Spark Migration Suite — 40+ SAS ETL scripts translated to PySpark Retail Analytics Databricks Pipeline — multi-source SQL → Delta ETL Healthcare KPI Engine — SAS logic re-implemented in Spark for 5× speedup I’ve done SAS → PySpark conversions many times and can guarantee exact match output.
$800 USD 7 gün içinde
5,4
5,4

✋ Hi there. I can convert your SAS data pipeline to PySpark on Azure Databricks, keeping all transformations and outputs identical while improving scalability and performance. ✔️ I have strong experience translating SAS workflows into PySpark, including data cleaning, transformations, and aggregations with SQL and Spark JDBC connections. I recently migrated a SAS ETL process to Databricks, maintaining row-level parity and faster runtimes, so I understand the requirements for a precise and maintainable pipeline. ✔️ For your project, I will rewrite the SAS logic in a well-structured Databricks notebook or .py files, connect to your SQL Database securely, and output the final tables with the same results. I will include clear inline comments so your team can maintain the code. ✔️ You will get the full notebook, verified outputs matching your SAS results, and a short hand-off session to review the process, run the pipeline, and confirm everything works as expected. Let’s chat so I can review your SAS files and plan the migration efficiently. Best regards, Mykhaylo
$500 USD 7 gün içinde
4,6
4,6

Drawing upon my extensive experience as a proficient Full Stack Developer and AI Specialist, I bring a unique set of skills that perfectly aligns with your project requirements. My fluency in Java and deep understanding of development conventions, combined with my ability to quickly adapt to new environments make me an ideal candidate for converting your SAS data pipeline to PySpark on Azure Databricks. I understand the importance of maintaining seamless, reliable processes when it comes to data migration, and my vast experience enables me to guarantee not only parity in transformation outputs but also performance improvements in PySpark. I can assure you that every line of code will adhere to PySpark and Databricks style conventions, enabling streamlined scheduling of the resulting job sans manual intervention. Further setting me apart is my AI integration expertise. Working with TensorFlow and PyTorch has provided me with valuable insights into handling large-scale data pipelines with precision and efficiency. I'm confident that I can successfully leverage this expertise to replicate your current SAS workflow while capitalizing on the scalability benefits Databricks has to offer.
$300 USD 7 gün içinde
4,2
4,2

Hello Dear! I write to introduce myself. I'm Engineer Toriqul Islam. I was born and grew up in Bangladesh. I speak and write in English like native people. I am a B.S.C. Engineer of Computer Science & Engineering. I completed my graduation from Rajshahi University of Engineering & Technology ( RUET). I love to work on Web Design & Development project. Web Design & development: I am a full-stack web developer with more than 10 years of experience. My design Approach is Always Modern and simple, which attracts people towards it. I have built websites for a wide variety of industries. I have worked with a lot of companies and built astonishing websites. All Clients have good reviews about me. Client Satisfaction is my first Priority. Technologies We Use: Custom Websites Development Using ======>Full Stack Development. 1. HTML5 2. CSS3 3. Bootstrap4 4. jQuery 5. JavaScript 6. Angular JS 7. React JS 8. Node JS 9. WordPress 10. PHP 11. Ruby on Rails 12. MYSQL 13. Laravel 14. .Net 15. CodeIgniter 16. React Native 17. SQL / MySQL 18. Mobile app development 19. Python 20. MongoDB What you'll get? • Fully Responsive Website on All Devices • Reusable Components • Quick response • Clean, tested and documented code • Completely met deadlines and requirements • Clear communication You are cordially welcome to discuss your project. Thank You! Best Regards, Toriqul Islam
$250 USD 5 gün içinde
4,0
4,0

Do you want the PySpark pipeline to include any additional logging or monitoring beyond what exists in the current SAS workflow? I will re-implement your entire SAS data processing pipeline in PySpark on Azure Databricks, ensuring all data cleaning, transformations, and aggregations produce outputs identical to your current SAS results. The deliverable will include a well-structured Databricks notebook (or Python scripts) with clear inline comments, secure SQL Database connections, and best-practice Spark JDBC code to pull from and write back to your database. I will also provide a short hand-off session to review the notebook, verify record-for-record parity with SAS outputs, and confirm the pipeline is ready for scheduled automated runs. Could you clarify if there are any performance benchmarks or maximum runtimes you’d like me to target in Databricks?
$600 USD 5 gün içinde
3,6
3,6

With over a decade of experience in the tech field and a flair for diverse coding languages, I believe I can help you transition from SAS data pipeline to PySpark seamlessly. My expertise includes Java and SQL which are vital in your project's execution. I have hands-on experience building and optimizing ETL processes using Java, significantly reducing runtime in projects, which I am confident will ensure that Databricks runs at the same pace or faster than SAS's batch. Working within multiple environments, I have cultivated the ability to adapt my skills to varied contexts while adhering to best practices, this will be invaluable as we leverage Azure Databricks' scalability for your project. Furthermore, having worked in relevant technologies such as Hadoop and Big Data Sales,I have the technical know-how of dealing with databases effectively and have no trouble establishing connections that ensure seamless data transfer between the SQL database and Databricks in line with industry-standard practices. Throughout my projects, delivering neat codes accompanied by clear in-line comments has been a top priority. Your team's future maintenance needs will be catered for as I document every step adequately. Finally, since I believe in quality work, our project hand-off won't be just a handover but an insightful session where we go through the full load together to guarantee all outputs match those of SAS on a record-for-record basis.
$500 USD 10 gün içinde
2,0
2,0

Hi, Requirement understanding: Need to migrate the current data flow from SAS into Databricks with proper comments and testing Experience: I haven’t worked with SAS earlier while I had migrated various data pipelines from other ETL tools like SSIS , SAP BW to Azure Databricks and consistent handover. Having total 15yrs of experience in data engineering universe using SQL , MSBI , ADF , Databricks etc I can help you to setup a right standards throughout pipeline from architect stand too. Let’s connect to discuss further
$750 USD 7 gün içinde
0,7
0,7

Hi, We went through your project description and it seems like our team is a great fit for this job. We are an expert team which have many years of experience on Java, Data Processing, SQL, SAS, Big Data Sales, Hadoop, ETL, PySpark Please come over chat and discuss your requirement in a detailed way. Thank You
$500 USD 7 gün içinde
0,5
0,5

Hi there, I’m excited about your project to convert your SAS data pipeline to PySpark on Azure Databricks. With extensive experience in translating complex SAS workflows to scalable PySpark solutions, I will ensure your data cleaning, transformations, and aggregation logic are faithfully replicated in a well-structured Databricks notebook. I will implement secure Spark JDBC connection code to integrate seamlessly with your SQL database, and provide clear inline documentation for easy maintenance. We’ll conclude with a comprehensive hand-off session to validate the outputs match exactly and ensure your team’s confidence. I suggest a timeline of 10-14 days to deliver a robust and efficient migration. Could you share the size and complexity of the dataset to help optimize the PySpark implementation? Thanks, Roshan
$550 USD 10 gün içinde
0,0
0,0

Hello, As an experienced C++ and Python developer specializing in full-stack web development, I've solved complex technical challenges and built high-performance systems, making me an ideal candidate for your project to convert the SAS data pipeline to PySpark. My 8 years of professional experience includes expertise in Java, which translates seamlessly to the PySpark ecosystem. Not only am I well-versed in the core languages and tools required for this task, but I'm also adept at working with databases using Spark JDBC. This means that not only will I be able to migrate your code effectively, but I'll also ensure secure connections to your SQL Database, enabling seamless data extraction and loading. Beyond the technical aspects, one crucial service I can offer is a clear hand-off session with your team. As so much rests on the successful migration of the existing logic into a new environment, it's crucial that your team understands the changes and can maintain the pipeline moving forward. My past experience includes ensuring clear documentation and comprehension on handovers, so this is an area where I excel. Thanks!
$450 USD 7 gün içinde
0,0
0,0

Hello there, I understand that you are looking to convert your existing SAS data pipeline to PySpark in order to leverage Azure Databricks for scalability and efficiency. Proposed Solution: I will re-implement your current SAS program in PySpark on Azure Databricks, ensuring a seamless transition of your data processing workflow. By closely following the provided .sas files and data dictionary, I will create a well-structured Databricks notebook that replicates the data cleaning, transformation, and aggregation logic. Additionally, I will establish a secure connection to the SQL Database for data extraction and loading, incorporating best practices for credentials and Spark JDBC usage. Key Deliverables: 1. Databricks notebook or .py files mirroring the SAS logic 2. Connection code for seamless SQL Database integration 3. Detailed in-line comments for easy maintenance 4. A comprehensive hand-off session for knowledge transfer and validation Portfolio & Skills: I bring extensive experience in translating SAS data pipelines to PySpark and working within Azure Databricks. I ensure adherence to best practices and efficient code execution for optimal performance. Call to Action: I'd love to discuss your project further and provide insights on how I can assist in this migration. Kindly reach out so we can schedule a chat to delve into the details. Best regards, Bilal
$500 USD 7 gün içinde
0,0
0,0

Hello there, I understand that you are looking to convert your existing SAS data pipeline to PySpark on Azure Databricks to improve scalability and efficiency. Recommended Approach: I plan to meticulously analyze the existing SAS program, replicate the data processing workflow in PySpark on Databricks, ensuring a seamless transition from SAS to PySpark. The focus will be on maintaining the integrity of the data transformations and ensuring optimal performance. Core Deliverables: - A well-structured Databricks notebook or .py files replicating the SAS logic for data cleaning, transformation, and aggregation. - Connection code for direct SQL Database access and writing final tables back using Spark JDBC. - Clear in-line comments for ease of maintenance. - A comprehensive hand-off session to review the notebook, validate outputs, and ensure parity with SAS results. Expertise & Portfolio: I'll share my portfolio with you in the DM. Kindly ping me there. My experience in translating SAS pipelines to PySpark ensures quality, consistency, and a smooth migration process. Invitation to Proceed: I'd be happy to discuss your project further and answer any questions you may have. Best regards, Malaika
$500 USD 7 gün içinde
0,0
0,0

Hi there, With 6+ years translating SAS ETL into PySpark pipelines on Azure Databricks, I will faithfully recreate your end-to-end SAS workflow in PySpark: data cleaning, transformations, and aggregations, then push outputs back to your SQL Database via Spark JDBC. The deliverable will be a well-structured Databricks notebook (or Python modules) with clear in-line comments, secure credential handling using Databricks secrets and Azure Key Vault, and a maintainable code style aligned with Databricks best practices. I’ll include a concise hand-off session to review the full load and verify parity against SAS outputs, ensuring all row counts and aggregations match within tolerance and performance meets or exceeds your current SAS batch. The solution will be designed for scheduling as a Databricks job with parameterization and robust error handling. 5-7 day turnaround is realistic given the provided artifacts (SAS code, data dictionary, and sample datasets). Best regards,
$450 USD 2 gün içinde
0,0
0,0

Hello Gopal R., I hope this message finds you well. I am excited about the opportunity to assist you in converting your SAS data pipeline to PySpark on Azure Databricks. With extensive experience in translating complex data workflows, I am confident in my ability to deliver a solution that meets your requirements. I understand that your current SAS program efficiently manages data processing for a SQL Database, and I will ensure that the new PySpark implementation mirrors this functionality. My approach will involve carefully analyzing the provided .sas files and data-dictionary to replicate the data cleaning, transformation, and aggregation steps in a well-structured Databricks notebook. I will also implement robust connection code to facilitate seamless data extraction and loading, adhering to best practices for security. In addition to clear, in-line comments for easy maintenance, I will conduct a thorough hand-off session to walk your team through the solution, ensuring all outputs match the existing SAS results. My goal is to achieve the required parity in row counts and aggregation values while optimizing runtime performance. I have successfully completed similar projects and can share relevant examples of my work, which demonstrate my commitment to quality and best practices. I look forward to the possibility of collaborating with you on this migration. Best regards, Marijo S.
$250 USD 2 gün içinde
0,0
0,0

Hi Gopal, I’ve reviewed your project details, and I’m confident I can convert your existing SAS data pipeline to PySpark efficiently. With extensive experience in transforming SAS workflows into PySpark environments, I understand the intricacies involved in data cleaning, transformation, and aggregation for SQL databases and can ensure a seamless transition to Azure Databricks. I’ll deliver a well-structured Databricks notebook or .py file that mirrors the SAS logic while providing robust connection code for optimal performance and security. Clear in-line comments will be included for future maintainability, and I’ll conduct a thorough hand-off session to validate that all outputs match your current SAS results. My goal is to ensure that the runtime in Databricks is not only comparable but ideally faster than your existing SAS setup, with meticulous attention to detail on all acceptance criteria. Please let me know if you have any specific preferences or questions.
$600 USD 7 gün içinde
0,0
0,0

Hi dear, hope you are doing well! With 9+ years of experience in software engineering and cloud data solutions, I have helped organizations modernize their data pipelines and seamlessly migrate legacy workflows to scalable cloud platforms like Azure Databricks. Gopal, your project is exciting because it combines SAS translation, PySpark engineering, and secure integration with SQL databases. I will carefully analyze your SAS code and data dictionary, then craft a clean, well-structured Databricks notebook (or .py files) that mirrors your existing logic end-to-end. I’ll ensure all data cleaning, transformation, and aggregation steps are faithfully reproduced, leveraging Spark best practices for optimal performance and maintainability. Secure JDBC integration will be set up for both reading and writing, with in-line comments throughout so your team feels confident maintaining the solution. After verifying that all outputs match perfectly with your SAS results, I’ll walk you through the pipeline and ensure everything is job-ready and fully automated. I value open communication and clear documentation, so you can always expect friendly updates, transparent collaboration, and a final result that’s both robust and easy to support. I’m genuinely excited to help you make this migration smooth and successful. Thanks
$500 USD 10 gün içinde
0,0
0,0

Hi, hope you are doing well. I've read your proposal very carefully, and I am confident about your project. I understand that you need to re-implement your existing SAS data processing workflow in PySpark on Azure Databricks, ensuring scalability and efficiency while maintaining the integrity of your outputs. I have hands-on experience in translating SAS workflows into PySpark, and I am well-versed in Azure Databricks environment. My approach includes: - Analyzing your existing SAS code and data-dictionary to ensure a seamless transition. - Developing a structured Databricks notebook that mirrors the SAS logic for data cleaning, transformation, and aggregation. - Implementing secure connection code for SQL Database access and ensuring that all outputs match the original SAS results. I can start immediately and complete the work within a short timeline while ensuring that all acceptance criteria are met. Looking forward to your reply.
$500 USD 7 gün içinde
0,0
0,0

Chantilly, United States
Ara 7, 2025 tarihinden bu yana üye
$30-250 USD
$250-750 USD
₹12500-37500 INR
₹750-1250 INR / saat
$250-750 CAD
₹12500-37500 INR
$30-250 USD
$30-250 USD
$30-250 USD
$250-750 USD
$250-750 USD
₹600-1500 INR
£20-250 GBP
$25-50 CAD / saat
$2-8 USD / saat
₹600-1500 INR
$10-30 USD
₹12500-37500 INR
₹100-400 INR / saat
$2-8 USD / saat
$250-750 USD