We are looking for an experienced Data Analyst to join our team on a long-term contract basis. This role is essential in ensuring the accuracy and integrity of client specialist data while maintaining high service standards. Based in Minneapolis, Minnesota, the position offers the opportunity to work remotely, collaborating across teams to deliver meaningful insights and solutions.<br><br>Responsibilities:<br>• Resolve data-related incidents by updating records and querying various data sources.<br>• Apply specialist quality standards to maintain data integrity.<br>• Validate data accuracy across multiple systems to ensure consistency.<br>• Update client data based on specific requirements while adhering to service level agreements.<br>• Configure software applications both before and after system go-live.<br>• Collaborate with care organizations to gather and update data in alignment with requirements.<br>• Provide support in managing data updates and ensuring the availability of accurate information.<br>• Utilize advanced Excel formulas to standardize and compare data efficiently.<br>• Foster effective communication with clients and stakeholders to address data needs and changes.
<p><strong>About the Role</strong></p><p>We’re seeking a Data Analyst to support reporting, analytics, and data-driven decision-making across the organization. This individual will extract, transform, analyze, and present data to business leaders, helping improve operations, customer experience, and strategic outcomes.</p><p><strong>Key Responsibilities</strong></p><ul><li>Collect, clean, and validate data from multiple systems (ERP, CRM, cloud platforms, internal databases).</li><li>Build dashboards, reports, and visualizations using tools such as Power BI, Tableau, or Excel.</li><li>Perform trend analysis, KPI tracking, and ad‑hoc analytics to support business decisions.</li><li>Partner with IT, finance, operations, and leadership teams to understand data needs and deliver insights.</li><li>Develop SQL queries, joins, stored procedures, and dataset pipelines for reporting.</li><li>Monitor data integrity and recommend process/data quality improvements.</li><li>Translate complex data findings into clear, actionable recommendations for non‑technical stakeholders.</li></ul><p><br></p>
We are looking for a highly skilled Data Analyst to join our team in Belgrade, Maine. This is a long-term contract position offering the opportunity to contribute to advanced data solutions and cloud-based projects. The role involves working with large-scale data migrations and developing modern data pipelines using cutting-edge technologies.<br><br>Responsibilities:<br>• Design and implement efficient data pipelines to ingest, transform, and deliver data across enterprise systems.<br>• Develop and manage Azure-based Data Lakes and associated cloud data services.<br>• Lead large-scale data migration projects from on-premise systems to Azure cloud environments.<br>• Create scalable data models and curated datasets to support business intelligence and reporting.<br>• Ensure data integrity, quality, governance, and security throughout all workflows.<br>• Collaborate with technical and business teams to translate requirements into impactful data solutions.<br>• Optimize data architectures to improve performance and scalability.<br>• Support business intelligence platforms like Power BI by delivering curated datasets and analytics layers.<br>• Monitor and troubleshoot data workflows to ensure seamless operations.
We are looking for a skilled Data Analyst to join our team on a long-term contract basis in Cincinnati, Ohio. In this role, you will leverage your expertise to analyze data patterns, identify potential fraudulent activities, and support investigations with actionable insights. This position offers an excellent opportunity to contribute to fraud prevention and detection strategies while working with diverse datasets.<br><br>Responsibilities:<br>• Analyze large datasets to identify trends, anomalies, and potential fraudulent activities.<br>• Develop and implement fraud detection models and analytics tools.<br>• Collaborate with fraud investigation teams to provide data-driven insights and recommendations.<br>• Monitor and assess data for suspected fraudulent behavior, ensuring timely identification and reporting.<br>• Create detailed reports and visualizations to support anti-fraud initiatives.<br>• Enhance existing fraud prevention strategies through continuous data analysis.<br>• Conduct regular audits of data systems to ensure accuracy and reliability.<br>• Work closely with cross-functional teams to improve fraud detection capabilities.<br>• Maintain and safeguard data integrity throughout all analytical processes.
<p>Overview</p><p>We are seeking a highly analytical Data Scientist / Advanced Analytics Specialist to leverage advanced analytics, machine learning, and statistical modeling techniques to extract insights from complex business data. This role focuses on building data‑driven models, conducting large‑scale experimentation, and influencing business decisions through actionable insights.</p><p>The ideal candidate is intellectually curious, thrives in ambiguous problem spaces, and enjoys applying cutting‑edge analytics techniques to real‑world business challenges.</p><p><br></p><p>Key Responsibilities</p><ul><li>Apply advanced analytics methods to extract value from structured and unstructured business data</li><li>Design and execute large‑scale experiments and develop data‑driven models to answer complex business questions</li><li>Conduct research on emerging techniques and tools in machine learning, deep learning, and artificial intelligence</li><li>Define requirements used to train, evaluate, and evolve predictive and deep learning models</li><li>Analyze results and present data‑based recommendations to product, business, and technical teams</li><li>Influence decision‑making through clear, compelling data storytelling and insights</li><li>Support ongoing analytics initiatives and perform additional duties as assigned</li></ul>
<p>Robert Half is seeking an experienced Data Architect to design and lead scalable, secure, and high-performing enterprise data solutions. This role will focus on building next-generation cloud data platforms, driving adoption of modern analytics technologies, and ensuring alignment with governance and security standards.</p><p><br></p><p>You’ll serve as a hands-on technical leader, partnering closely with engineering, analytics, and business teams to architect data platforms that enable advanced analytics and AI/ML initiatives. This position blends deep technical expertise with strategic thinking to help unlock the value of data across the organization.</p><p><br></p><p><strong>Key Responsibilities:</strong></p><ul><li>Design and implement end-to-end data architecture for big data and advanced analytics platforms.</li><li>Architect and build Delta Lake–based lakehouse environments from the ground up, including DLT pipelines, PySpark jobs, workflows, Unity Catalog, and Medallion architecture.</li><li>Develop scalable data models that meet performance, security, and governance requirements.</li><li>Configure and optimize clusters, notebooks, and workflows to support ETL/ELT pipelines.</li><li>Integrate cloud data platforms with supporting services such as data storage, orchestration, secrets management, and analytics tools.</li><li>Establish and enforce best practices for data governance, security, and cost optimization.</li><li>Collaborate with data engineers, analysts, and stakeholders to translate business requirements into technical solutions.</li><li>Provide technical leadership and mentorship to team members.</li><li>Monitor, troubleshoot, and optimize data pipelines to ensure reliability and efficiency.</li><li>Ensure compliance with organizational and regulatory standards related to data privacy and security.</li><li>Create and maintain documentation for architecture, processes, and governance standards.</li></ul><p><br></p>
<p>We are looking for an experienced Data Architect to join our team on a long-term contract basis in Cleveland, Ohio. This role involves designing scalable enterprise data platforms, ensuring data quality, and implementing robust data governance frameworks. You will play a pivotal role in leveraging Azure services and AI-driven analytics to optimize data architecture and enhance operational insights.</p><p><br></p><p>Responsibilities:</p><p>• Develop and implement enterprise-wide data architectures and canonical data models.</p><p>• Establish data ownership protocols, governance standards, and quality benchmarks.</p><p>• Analyze and stabilize data pipelines across distributed systems and platforms.</p><p>• Perform detailed data analysis and reconciliation to identify and resolve integrity issues.</p><p>• Design and implement monitoring tools to validate and improve data quality.</p><p>• Enhance data observability and lineage tracking to streamline governance processes.</p><p>• Utilize AI-driven analytics and automation to detect anomalies and accelerate decision-making.</p><p>• Collaborate with engineering teams to align data architecture with integration services and platform requirements.</p><p>• Optimize event-driven and distributed data systems for scalability and reliability.</p><p>• Conduct hands-on work with Azure services, such as Azure Data Factory and Synapse, to implement solutions.</p>
<p><strong>Senior Data Scientist</strong></p><p>Hybrid | Austin, TX | Contract</p><p><br></p><p>Robert Half is partnering with an Austin-based client to hire a Senior Data Scientist for a long-term contract. Applicants must have a PhD in Statistics, Economics, or related quantitative discipline.</p><p><br></p><p><strong>Responsibilities:</strong></p><ul><li>Lead the design, execution, and interpretation of A/B tests and quasi-experiments to evaluate user migration initiatives, growth marketing efforts, and campaign effectiveness.</li><li>Partner closely with cross-functional teams across product, engineering, and marketing to embed experimentation into product development and iteration cycles.</li><li>Serve as a subject-matter expert on experimentation best practices, including hypothesis formulation, metric selection, experimental design, and results interpretation.</li><li>Apply advanced causal inference techniques when randomized experiments are not feasible, or to inform prioritization and design.</li><li>Contribute to the development and scaling of centralized experimentation frameworks, tools, and documentation across the organization.</li><li>Independently extract, transform, and analyze data from complex, large-scale systems using SQL, Python, and related analytics tools.</li><li>Communicate insights and experimental results clearly and effectively to both technical and non-technical stakeholders to drive informed business decisions.</li><li>Stay current on emerging methodologies in experimentation, causal analysis, and applied statistics, and bring new ideas into practice.</li></ul>
We are looking for an experienced Senior Data Scientist to join our dynamic team in Boston, Massachusetts. In this role, you will leverage your expertise in statistical modeling, machine learning, and cloud-based analytics to drive impactful decisions and solutions. The ideal candidate will bring a strong technical background, a passion for working with regulated data, and a commitment to ethical AI practices.<br><br>Responsibilities:<br>• Develop and implement advanced statistical models and machine learning algorithms to solve complex business problems.<br>• Monitor and evaluate the performance of AI models, ensuring reliability, fairness, and compliance with ethical standards.<br>• Collaborate with engineering and product teams to translate data-driven insights into actionable strategies.<br>• Utilize cloud-based tools such as AWS SageMaker and Redshift to design and deploy scalable analytics solutions.<br>• Handle sensitive healthcare or clinical trial datasets while adhering to strict data privacy and security regulations.<br>• Conduct exploratory data analysis and create visualizations to communicate findings effectively.<br>• Build and optimize ETL pipelines for efficient data transformation and integration.<br>• Apply Bayesian statistics and time-series forecasting techniques to improve predictive accuracy.<br>• Maintain comprehensive documentation of data science workflows and processes.<br>• Stay updated on industry trends and advancements to continuously enhance methodologies and tools.
<p><strong>Data Modeling and Analysis</strong></p><ul><li>Design data models and optimize performance: Creating the structure of data relationships ensuring efficient data retrieval and calculations.</li><li>Create calculated columns and measures: Using DAX to calculate derived values and aggregate metrics.</li><li>Perform exploratory data analysis (EDA): Using BI tools to explore data, identify trends, and patterns.</li><li>Apply advanced data analysis techniques (e.g., statistical analysis, time series analysis, predictive modeling).</li><li>Integrate machine learning models into Power BI dashboards.</li><li>Experience building semantic models</li></ul><p><strong>Dashboard Development and Visualization</strong></p><ul><li>Designing dashboards: Creating visually appealing and interactive dashboards.</li><li>Creating visualizations: Using charts, graphs, and other visual elements to represent data.</li><li>Implementing interactivity: Adding filters, slicers, and drill-down capabilities.</li><li>Expertise in SQL and DAX and knowledge of Python, R.</li><li>Strong proficiency in Power BI.</li><li>Data modeling and visualization skills.</li><li>Strong problem-solving skills to address technical challenges and data quality issues.</li><li>Analytical skills with capacity to analyze complex data problems and draw meaningful insights.</li></ul>
We are looking for a skilled Data Engineer to join our team in Foxborough, Massachusetts, on a long-term contract basis. In this role, you will design, optimize, and maintain data pipelines and storage solutions, leveraging modern tools to ensure high performance and reliability. This position offers an exciting opportunity to collaborate across teams and implement cutting-edge practices in data engineering and analytics.<br><br>Responsibilities:<br>• Optimize Amazon Redshift performance by configuring distribution keys, sort keys, and fine-tuning queries.<br>• Develop and maintain robust data pipelines using AWS Glue and orchestrate workflows with Airflow.<br>• Manage semantic layers and metadata to support reliable analytics and AI-driven insights.<br>• Implement best practices for data partitioning, compression, and columnar storage formats.<br>• Monitor and troubleshoot data workflows to ensure high availability, reliability, and automated observability.<br>• Automate data processing tasks using Python and AWS native tools.<br>• Enforce data security and governance policies, including row- and column-level controls, using Lake Formation and AWS services.<br>• Oversee compliance monitoring and auditing through CloudWatch, CloudTrail, and similar tools.<br>• Continuously refine and improve data architecture by adopting emerging AWS best practices and patterns.<br>• Collaborate closely with Operations, Data Governance, and other teams to align with standards and achieve delivery objectives.
We are looking for a Senior Data Engineer to develop and optimize enterprise data systems that support analytics and digital solutions. In this role, you will design and implement robust data architectures, ensuring seamless data integration and transformation processes across the organization. Your expertise will drive the creation of reliable pipelines and scalable infrastructure, enabling advanced analytics and machine learning capabilities.<br><br>Responsibilities:<br>• Design and implement scalable data pipelines using Databricks, Spark, and Delta Lake to support enterprise-level analytics.<br>• Develop and maintain efficient data models tailored for AI, analytics, and operational systems.<br>• Lead Master Data Management initiatives to establish unified and accurate data records across platforms.<br>• Create batch and near-real-time data processing workflows for structured and semi-structured datasets.<br>• Collaborate with AI and software development teams to ensure delivery of high-quality datasets for machine learning.<br>• Define and enforce data architecture standards, ensuring scalability, reliability, and governance.<br>• Troubleshoot and optimize data systems to maintain performance and reliability in complex environments.<br>• Partner with cloud and IT teams to integrate modern data platforms and ensure seamless functionality.
<p>Our transportation client is seeking a <strong>Data Engineer</strong> to support large‑scale logistics operations by building reliable, scalable, and cloud‑based data pipelines. This role is hands‑on, focused on delivering high‑quality data flows that improve shipment visibility, operational efficiency, and real‑time analytics across the supply chain.</p><p><br></p><p><strong>Key Responsibilities</strong></p><ul><li>Design, build, and maintain <strong>ETL/ELT pipelines</strong> that process high‑volume operational and logistics data</li><li>Develop transformation logic and automation using <strong>Python</strong>, <strong>SQL</strong>, and Azure-native tooling</li><li>Implement and orchestrate workflows in <strong>Azure Data Factory</strong>, <strong>Synapse</strong>, and <strong>Databricks</strong></li><li>Optimize data lake and warehouse performance, including tuning queries, pipelines, and storage layers</li><li>Monitor pipeline health and proactively troubleshoot failures, bottlenecks, and data quality issues</li><li>Contribute to data modeling efforts to support analytics, reporting, and downstream applications</li><li>Collaborate with BI, product, supply chain, and application teams to align pipelines with business needs</li><li>Maintain strong documentation around workflows, standards, and operational procedures</li><li>Support governance initiatives related to <strong>data quality</strong>, lineage, cataloging, and access policies</li><li>Follow best practices for security, compliance, and cloud resource management</li></ul><p><br></p><p><br></p>
We are looking for an experienced Data Engineer to join our team in Cincinnati, Ohio. This long-term contract position offers the opportunity to work on cutting-edge data engineering projects while collaborating with multidisciplinary teams to deliver high-quality solutions. The ideal candidate will have a strong background in Databricks and big data technologies, along with a passion for optimizing data processes and systems.<br><br>Responsibilities:<br>• Design, build, and enhance data pipelines using Databricks Runtime, Delta Lake, Autoloader, and Structured Streaming.<br>• Implement secure and governed data access protocols utilizing Unity Catalog, workspace controls, and audit configurations.<br>• Manage and integrate structured and unstructured data from diverse sources, including APIs and cloud storage.<br>• Develop and maintain notebook-based workflows and manage jobs using Databricks Workflows and Jobs.<br>• Apply best practices for performance tuning, scalability, and cost optimization in Databricks environments.<br>• Collaborate with data scientists, analysts, and business stakeholders to deliver clean and reliable datasets.<br>• Support continuous integration and deployment processes for Databricks jobs and system configurations.<br>• Ensure high standards of data quality and security across all engineering tasks.<br>• Troubleshoot and resolve issues to maintain operational efficiency in data pipelines.
<p>We are seeking a Data Engineer to design, build, and maintain scalable data pipelines and infrastructure. This role will support data-driven decision-making by ensuring reliable data flow, transformation, and accessibility across the organization.</p><p><br></p><p>Key Responsibilities</p><ul><li>Design, build, and maintain ETL/ELT data pipelines</li><li>Develop and optimize data models and data architectures</li><li>Integrate data from multiple sources (APIs, databases, third-party systems)</li><li>Ensure data quality, integrity, and reliability</li><li>Collaborate with data analysts, data scientists, and business stakeholders</li><li>Monitor and troubleshoot data pipeline performance issues</li><li>Implement best practices for data governance and security</li></ul><p><br></p>
<p>We are looking for a talented Data Engineer to join our team in Fort Lauderdale, Florida. This long-term contract position offers the opportunity to work on cutting-edge technologies and contribute to the development of efficient data pipelines and processes. The ideal candidate will have a strong background in data engineering and a passion for delivering high-quality solutions that drive business success.</p><p><br></p><p>Responsibilities:</p><p>• Design and implement scalable data pipelines using Snowflake, Python, and other relevant tools.</p><p>• Collaborate with stakeholders to gather and refine data requirements, ensuring alignment with business needs.</p><p>• Develop and maintain data models to support analytics, reporting, and operational processes.</p><p>• Optimize data warehouse performance by tuning queries and managing resources effectively.</p><p>• Ensure data quality through rigorous testing and governance protocols.</p><p>• Implement security and compliance measures to protect sensitive data.</p><p>• Research and integrate emerging technologies to enhance system capabilities.</p><p>• Support ETL processes for data extraction, transformation, and loading.</p><p>• Work with technologies such as Apache Spark, Hadoop, and Kafka to manage and process large datasets.</p><p>• Provide technical guidance and support to team members and stakeholders.</p>
We are looking for an experienced Data Engineer to join our team on a long-term contract basis. Based in Houston, Texas, this role offers an exciting opportunity to work with cutting-edge data technologies, design scalable solutions, and contribute to data-driven decision-making processes. If you are passionate about optimizing data systems and driving innovation, we encourage you to apply.<br><br>Responsibilities:<br>• Develop, maintain, and optimize scalable data pipelines using Apache Spark and Python.<br>• Implement ETL processes to ensure seamless extraction, transformation, and loading of data across systems.<br>• Collaborate with cross-functional teams to integrate Apache Hadoop and Apache Kafka into the data architecture.<br>• Monitor and troubleshoot data systems to ensure reliability and performance.<br>• Design and maintain data models, ensuring alignment with business requirements.<br>• Conduct thorough testing and validation of data processes to guarantee accuracy.<br>• Document data workflows and processes for future reference and team collaboration.<br>• Provide technical guidance and support to team members on data engineering best practices.<br>• Stay current on emerging technologies and trends in big data and analytics.<br>• Contribute to improving data governance and security protocols.
We are looking for a skilled Data Engineer to join our team in Houston, Texas. This Contract to permanent position offers an exciting opportunity to work at the intersection of data engineering, analytics, and business strategy. If you have a strong background in building and optimizing data pipelines and are passionate about leveraging technology to drive insights, we encourage you to apply.<br><br>Responsibilities:<br>• Design, develop, and optimize scalable data pipelines and workflows to support business analytics.<br>• Collaborate with cross-functional teams to gather and analyze data requirements.<br>• Implement ETL processes to extract, transform, and load data from diverse sources.<br>• Utilize tools such as Apache Spark and Hadoop to manage large-scale data processing.<br>• Integrate streaming data systems using Apache Kafka to enhance real-time analytics.<br>• Monitor and troubleshoot data flow and systems to ensure high performance and reliability.<br>• Develop and maintain documentation for data engineering processes and systems.<br>• Ensure data security and integrity across all platforms and processes.<br>• Work closely with stakeholders to translate business needs into technical solutions.<br>• Stay updated with industry trends and emerging technologies to improve data engineering practices.
<p>I’m building a world-class team to power our next generation of data products. We’re looking for a Senior Data Engineer who knows AWS inside and out—someone who can <strong>design secure, scalable data pipelines</strong>, <strong>own ETL/ELT workflows</strong>, <strong>engineer cloud data infrastructure</strong>, and <strong>deliver dimensional and semantic models</strong> that our analysts, data scientists, and applications can trust.</p><p>You’ll work closely with product, security, platform engineering, and analytics to move our architecture toward a <strong>real-time, governed, cost-aware</strong>, and <strong>highly automated</strong> data ecosystem.</p><p><strong>What You’ll Do</strong></p><ul><li><strong>Design & build end-to-end pipelines</strong> on AWS (batch and streaming) using services like <strong>Glue, EMR, Lambda, Step Functions, Kinesis, MSK</strong>, and <strong>Fargate</strong>.</li><li><strong>Develop robust ETL/ELT</strong> (PySpark, Spark SQL, SQL, Python) for structured, semi-structured, and unstructured data at scale.</li><li><strong>Own data storage & processing layers</strong>: <strong>S3 (Lake/Lakehouse), Redshift (or Snowflake on AWS), DynamoDB</strong>, and <strong>Athena</strong> with strong partitioning, compaction, and performance tuning.</li><li><strong>Implement data models</strong> (3NF, dimensional/star, Data Vault, Lakehouse medallion) for analytics and operational workloads.</li><li><strong>Engineer secure infrastructure-as-code</strong> with <strong>Terraform</strong> (or <strong>CDK</strong>) across multi-account setups; implement CI/CD via <strong>GitHub Actions</strong> or <strong>AWS CodeBuild/CodePipeline</strong>.</li><li><strong>Harden security & governance</strong>: use <strong>IAM</strong>, <strong>Lake Formation</strong>, <strong>KMS</strong>, <strong>Secrets Manager</strong>, <strong>VPC/PrivateLink</strong>, <strong>GLUE Catalog</strong>, and fine-grained access controls. Partner with SecOps on compliance (e.g., <strong>SOC 2</strong>, <strong>FedRAMP</strong>, <strong>HIPAA</strong> depending on dataset).</li><li><strong>Observability & reliability</strong>: build monitoring with <strong>CloudWatch</strong>, <strong>OpenTelemetry</strong>, and data quality checks (e.g., <strong>Great Expectations</strong>, <strong>Deequ</strong>), implement SLOs and alerts.</li><li><strong>Champion best practices</strong>: code reviews, testing (unit/integration), documentation, runbooks, and blameless postmortems.</li><li><strong>Mentor</strong> mid-level engineers and collaborate on architectural decisions, standards, and technical roadmaps.</li></ul><p><br></p>
<p><strong>Overview</strong></p><p>We are looking for a <strong>Data Engineer </strong>to design, build, and maintain data solutions that enable reporting, analytics, and informed decision‑making.</p><p><strong>Responsibilities</strong></p><ul><li>Design and maintain data pipelines and data models</li><li>Extract, transform, and load (ETL) data from multiple sources</li><li>Develop dashboards, reports, and analytics for business users</li><li>Ensure data accuracy, integrity, and governance</li><li>Collaborate with stakeholders to understand reporting needs</li></ul><p><br></p>
We are looking for a talented Data Engineer to join our team in Grand Rapids, Michigan. In this role, you will focus on designing, building, and optimizing robust data solutions using Snowflake and other cloud-based technologies. You will work closely with business intelligence and analytics teams to deliver scalable, high-performance data pipelines that support organizational goals.<br><br>Responsibilities:<br>• Design and implement scalable data models, schemas, and tables within Snowflake, including staging, integration, and presentation layers.<br>• Develop and optimize data pipelines using Snowflake tools such as Snowpipe, Streams, Tasks, and stored procedures.<br>• Ensure data security and access through role-based controls and best practices for data sharing.<br>• Build and maintain ETL pipelines leveraging tools like dbt, Matillion, Fivetran, Informatica, or Azure-native solutions.<br>• Integrate data from diverse sources such as APIs, IoT devices, and NoSQL databases to create unified datasets.<br>• Enhance performance by utilizing clustering, partitioning, caching, and efficient warehouse sizing strategies.<br>• Collaborate with cloud technologies such as AWS, Azure, or Google Cloud to support Snowflake infrastructure and operations.<br>• Implement automated workflows and CI/CD processes for seamless deployment of data solutions.<br>• Maintain high standards for data accuracy, completeness, and reliability while supporting governance and documentation.<br>• Work closely with analytics, reporting, and business teams to troubleshoot issues and deliver scalable solutions.
<p>We are looking for an experienced Data Engineer to join our team on a contract basis in Columbus, Ohio. In this role, you will take on a leadership position, driving the development and optimization of data pipelines that support enterprise-wide analytics and decision-making. You will also play a key role in mentoring team members, fostering collaboration, and ensuring the integrity and quality of data across various business functions.</p><p><br></p><p>Responsibilities:</p><p>• Design, develop, and maintain efficient data pipelines to support enterprise analytics and reporting.</p><p>• Collaborate with business analysts and data science teams to refine data requirements and ensure alignment with organizational goals.</p><p>• Enhance and automate data integration and management processes to improve operational efficiency.</p><p>• Lead efforts to ensure data quality by testing for accuracy, consistency, and conformity to business rules.</p><p>• Provide training and guidance to team members and other stakeholders on data pipelining and preparation techniques.</p><p>• Partner with data governance teams to promote vetted content into the curated data catalog for reuse.</p><p>• Stay updated on emerging technologies and assess their impact on current systems and processes.</p><p>• Offer leadership, coaching, and mentorship to team members, encouraging attention to detail in their development.</p><p>• Work closely with stakeholders to understand business needs and ensure solutions meet those requirements.</p><p>• Perform additional duties as assigned to support organizational objectives.</p>
<p>We are looking for a talented Data Engineer to join our team in Miami, Florida. This long-term contract position offers the opportunity to work on cutting-edge technologies and contribute to the development of efficient data pipelines and processes. The ideal candidate will have a strong background in data engineering and a passion for delivering high-quality solutions that drive business success.</p><p><br></p><p>Responsibilities:</p><p>• Design and implement scalable data pipelines using Snowflake, Python, and other relevant tools.</p><p>• Collaborate with stakeholders to gather and refine data requirements, ensuring alignment with business needs.</p><p>• Develop and maintain data models to support analytics, reporting, and operational processes.</p><p>• Optimize data warehouse performance by tuning queries and managing resources effectively.</p><p>• Ensure data quality through rigorous testing and governance protocols.</p><p>• Implement security and compliance measures to protect sensitive data.</p><p>• Research and integrate emerging technologies to enhance system capabilities.</p><p>• Support ETL processes for data extraction, transformation, and loading.</p><p>• Work with technologies such as Apache Spark, Hadoop, and Kafka to manage and process large datasets.</p><p>• Provide technical guidance and support to team members and stakeholders.</p>
<p>We are seeking a highly skilled Data Engineer to design, build, and manage our data infrastructure. The ideal candidate is an expert in writing complex SQL queries, designing efficient database schemas, and developing ETL/ELT pipelines. This role ensures data accuracy, accessibility, and performance optimization to support business intelligence, analytics, and reporting initiatives.</p><p><br></p><p><strong><em><u>Key Responsibilities</u></em></strong></p><p><br></p><p><strong>Database Design & Management</strong></p><ul><li>Design, develop, and maintain relational databases, including SQL Server, PostgreSQL, and Oracle, as well as cloud-based data warehouses.</li></ul><p><strong>Strategic SQL & Data Engineering</strong></p><ul><li>Develop advanced, optimized SQL queries, stored procedures, and functions to process and analyze large, complex datasets and deliver actionable business insights.</li></ul><p><strong>Data Pipeline Automation & Orchestration</strong></p><ul><li>Build, automate, and orchestrate ETL/ELT workflows using SQL, Python, and cloud-native tools to integrate and transform data from diverse, distributed sources.</li></ul><p><strong>Performance Optimization</strong></p><ul><li>Tune SQL queries and optimize database schemas through indexing, partitioning, and normalization to improve data retrieval and processing performance.</li></ul><p><strong>Data Integrity & Security</strong></p><ul><li>Ensure data quality, consistency, and integrity across systems.</li><li>Implement data masking, encryption, and role-based access control (RBAC).</li></ul><p><strong>Documentation</strong></p><ul><li>Maintain comprehensive technical documentation, including database schemas, data dictionaries, and ETL workflows.</li></ul>
<p>Robert Half is hiring! We are looking for an experienced Data Engineer to join our team in Greenville, South Carolina. This role offers an exciting opportunity to work with modern data technologies, ensuring the efficient operation and optimization of data pipelines and systems. The ideal candidate will bring a strong technical background, leadership skills, and a proactive approach to maintaining and improving data infrastructure.</p><p><br></p><p>Responsibilities:</p><p>• Oversee daily data loads and ensure the smooth operation of data pipelines and related systems.</p><p>• Troubleshoot and resolve issues such as pipeline failures, performance bottlenecks, schema mismatches, and cloud resource disruptions.</p><p>• Conduct root-cause analyses and implement permanent solutions to prevent recurring issues.</p><p>• Maintain and optimize existing data processes, refactoring or retiring outdated workflows as necessary.</p><p>• Design and build scalable data ingestion pipelines using technologies such as Azure Data Factory, Databricks, and Synapse Pipelines.</p><p>• Collaborate with teams to create and improve operational runbooks, monitoring dashboards, and incident response workflows.</p><p>• Develop reusable ingestion patterns for platforms like Guidewire DataHub, InfoCenter, and other business data sources.</p><p>• Lead the implementation of real-time and event-driven data engineering solutions to enable operational insights and automation.</p><p>• Partner with architects to modernize data workloads using advanced frameworks like Delta Lake and Medallion Architecture.</p><p>• Mentor entry-level engineers, enforce coding best practices, and review code to ensure quality and compliance.</p>