Data Engineer<p>We are offering a short-term contract employment opportunity for a Data Engineer in Chicago, Illinois. As a Data Engineer, you will be instrumental in developing new data pipelines, enhancing current data architecture, and troubleshooting any issues that may arise. Your role will also involve implementing Azure data architectures and environments and collaborating within a project team to solve complex problems.</p><p><br></p><p>Responsibilities:</p><p>• Develop and architect new data pipelines to handle large amounts of data</p><p>• Enhance and optimize current data pipelines and data architecture to meet emerging data and analytic needs</p><p>• Troubleshoot and resolve issues and bottlenecks within the current data architecture</p><p>• Implement and design Azure data architectures and environments, optimizing applications for Azure when necessary</p><p>• Collaborate within a project team environment to solve complex problems, contributing to project success</p><p>• Utilize Azure services for data migration and data processing, including Serverless Architecture, Azure Storage, Azure SQL DB/DW, Data Factory, and more</p><p>• Maintain familiarity with the technology stack available for data management, data ingestion, capture, processing, and curation</p><p>• Work with cloud migration methodologies and processes, including tools like Azure Data Factory, Event Hub, etc.</p><p>• Display a strong knowledge of private and public cloud architectures, their pros/cons, and migration considerations.</p><p>• Contribute to a team-oriented environment, fostering creativity and problem-solving skills.</p>Data EngineerWe are seeking a Data Engineer to join our team based in Cleveland, Ohio. This role primarily involves working with Azure technologies including Azure Data Factory, Azure Data Lake, and Azure Synapse Analytics. The Data Engineer will be responsible for managing and optimizing data pipelines and developing techniques for big data handling and analysis.<br><br>Responsibilities:<br>• Design, construct, install, test and maintain data management systems.<br>• Ensure systems meet business requirements and industry practices.<br>• Build high-performance algorithms, prototypes, and conceptual models.<br>• Identify ways to improve data reliability, efficiency, and quality.<br>• Collaborate with data architects, modelers, and IT team members on project goals.<br>• Use Azure Data Factory for the orchestration and ETL/ELT data movement.<br>• Utilize Azure Data Lake for storing and analyzing data.<br>• Implement Azure Synapse Analytics to blend big data and relational data.<br>• Translate business needs into data solutions.<br>• Conduct testing and troubleshooting of data systems.Data Engineer<p>We are inviting applications for a Data Engineer role in Fairfield County, Connecticut, the successful candidate will play a pivotal role in enhancing our data infrastructure and pipelines, ensuring data accuracy and reliability.</p><p><br></p><p>Responsibilities:</p><p>• Design, develop, and maintain our data infrastructure and pipelines to ensure efficient data processing and storage.</p><p>• Incorporate processes and systems for monitoring data quality, ensuring that the data needed for key stakeholders and business processes is always accurate and available.</p><p>• Identify and resolve data inconsistencies, errors, and anomalies to ensure data quality and reliability.</p><p>• Collaborate with data scientists, analysts, and other stakeholders to comprehend data needs and requirements.</p><p>• Enhance internal processes by automating manual processes, optimizing data delivery, and re-designing infrastructure for better scalability.</p><p>• Monitor and optimize data performance, troubleshooting, and resolving issues as they arise.</p><p>• Stay updated with new technologies and approaches in data engineering, suggesting improvements to our data infrastructure.</p><p>• Handle structured, semi-structured, and unstructured data sources effectively.</p><p>• Implement cloud-native data pipeline architectures using Azure or AWS.</p><p>• Utilize SQL and Python to manage large data sets.</p>Data EngineerWe are seeking a Data Engineer to join our team. As a Data Engineer, you will be responsible for designing and implementing data engineering solutions, leveraging tools like Snowflake, DBT, and Fivetran. This role offers a long term contract employment opportunity. <br><br>Responsibilities:<br>• Design, develop, and maintain data engineering solutions using Snowflake<br>• Implement ETL processes using DBT and other tools<br>• Utilize Fivetran for data ingestion tasks<br>• Ensure the efficient scheduling of data warehouse tasks with tools like Airflow or Dagster<br>• Develop and maintain APIs for data interaction<br>• Implement algorithms and analytics for data processing<br>• Leverage Apache Kafka, Apache Pig, and Apache Spark in data processing tasks<br>• Use AWS technologies for cloud-based data engineering solutions<br>• Implement data visualization techniques for data presentation<br>• Mentor entry level team members on data warehouse concepts and practices<br>• Ensure clear and effective communication with team members and stakeholders.Data EngineerWe are offering a long term contract employment opportunity for a proficient Data Engineer in Pittsburgh, Pennsylvania. The selected candidate will be primarily tasked with converting data from legacy Microsoft tools to modern AWS services, employing their expertise within the federal government contractor industry.<br><br>Responsibilities:<br><br>• Execute data migration from legacy Microsoft tools to AWS services with a focus on maintaining data integrity and accuracy.<br>• Utilize data visualization tools like Tableau or Power BI to process, analyze, and present data during migration workflows.<br>• Build and manage comprehensive ETL pipelines using AWS Glue, Athena, and related AWS tools.<br>• Employ AWS Step Functions to automate workflows and synchronize distributed application services.<br>• Establish secure data storage, processing, and access workflows in AWS S3 and RDS.<br>• Leverage AWS Kinesis for real-time data streaming in scalable data migration strategies.<br>• Collaborate with cross-functional teams to ensure all data migrations adhere to federal compliance standards and organizational requirements.<br>• Provide technical expertise on federal regulatory constraints, security policies, and operational best practices unique to the AWS.gov environment.<br>• Troubleshoot and resolve migration-related issues while mitigating risks during the transition process.<br>• Document processes, technical configurations, and migration steps effectively for audits and team reference.Data Engineer<p>We are seeking an experienced Azure Data Engineer to join our growing team. The successful candidate will play a pivotal role in designing, implementing, and optimizing data solutions on Microsoft Azure, ensuring seamless integration, scalability, and performance. You will work closely with cross-functional teams to deliver high-quality data infrastructure and analytics solutions for our organization and clients.</p><p>Key Responsibilities</p><ul><li>Design and implement data pipelines, ETL/ELT processes, and workflows using Azure Data Factory and related tools.</li><li>Develop and maintain data storage solutions, including Azure Data Lake, Azure SQL Database, and Azure Synapse Analytics.</li><li>Ensure data integrity, security, and performance optimization for all deployed solutions.</li><li>Collaborate with stakeholders to understand data requirements and translate them into scalable architectures.</li><li>Utilize Azure Databricks or other Big Data platforms for advanced analytics and modeling workflows.</li><li>Implement and monitor data processing systems, ensuring reliability and fault tolerance.</li><li>Integrate on-premises and cloud data sources seamlessly with Azure platforms.</li><li>Work on Data Governance initiatives, ensuring compliance with organizational standards and policies.</li></ul><p><br></p>Data EngineerWe are offering a long-term contract employment opportunity for a Data Engineer in OAKWOOD VILLAGE, Ohio. As a Data Engineer, you will be responsible for implementing algorithms, developing APIs, and utilizing various technologies to analyze and visualize data. You will also play a crucial role in extracting, reporting, and warehousing data.<br><br>Responsibilities:<br>• Utilizing Apache Kafka, Apache Pig, and Apache Spark to process and analyze data<br>• Implementing and optimizing cloud technologies for data storage and analysis<br>• Creating and maintaining data visualizations using Microsoft Power BI and Excel<br>• Implementing algorithms for complex data analysis and processing<br>• Developing and maintaining analytics using Apache Hadoop and AWS technologies<br>• Extracting relevant data from various sources for analysis and reporting<br>• Developing and maintaining data lakes for efficient data storage and retrieval<br>• Reporting data in a clear and understandable manner for stakeholders<br>• Ensuring the accuracy and integrity of data through data warehousing.Data EngineerWe are operating in the Aerospace industry, situated in SWANTON, Ohio, United States, and we are on the lookout for a Data Engineer. The selected candidate will design and support our data systems, carry out maintenance processes, and manage projects using Azure services such as Data Factory, Databricks, Synapse Analytics, Data Lake, ETL, and Microsoft Fabric. The role also demands proficiency in SQL programming languages like Python or Scala.<br><br>Responsibilities:<br><br>• Design and support data systems using Azure services such as Data Factory, Databricks, Synapse Analytics, Data Lake, ETL, and Microsoft Fabric.<br>• Execute data implementations following system requirements and anticipated usage.<br>• Conduct data extraction operations while ensuring data accuracy.<br>• Troubleshoot and resolve issues during data operations.<br>• Analyze business intelligence data, design, and generate relevant reports.<br>• Utilize ETL tools and programming languages effectively.<br>• Act as a liaison between IT and business units, facilitating efficient communication and collaboration.<br>• Ensure the proficiency of SQL programming languages like Python or Scala.<br>• Maintain blob storage and oversee its efficient functioning.<br>• Play an active role in maintaining the accuracy of customer credit records.Data EngineerWe are offering a permanent employment opportunity for a Data Engineer in the insurance industry, based in Branchville, New Jersey. As a Data Engineer, your role will encompass the development and support of data applications, working in close collaboration with business analysts, application and enterprise architects, and providing technical guidance to the team for implementing complex data solutions.<br><br>Responsibilities:<br>• Collaborate with business analysts to understand data and business processes, making recommendations on best practices or long-term solutions to resolve current issues and for future system design.<br>• Engage in hands-on development and support of new or existing data applications.<br>• Work in partnership with Application and Enterprise Architects to understand high-level data flow designs and create/review low-level implementation designs.<br>• Offer technical guidance to the team for implementing intricate data solutions.<br>• Maintain detailed documentation to support downstream integrations.<br>• Provide support for production issues and execute the activities of a scrum master.<br>• Identify technology trends and explore opportunities for their use within the organization. <br>• Participate in the design, development, code reviews, testing, deployment, and documentation of data engineering and data integration applications.Data Engineer<p>We are seeking a highly skilled and experienced Data Engineer to join our growing data team. This role will be pivotal in building and maintaining our data infrastructure, focusing on leveraging Azure cloud technologies and Snowflake. The ideal candidate will have at least 5 years of experience in data engineering, with a strong background in Azure Data Factory, Azure Synapse, Snowflake, Python, Java, and CI/CD pipelines.</p><p><br></p><p><strong>Responsibilities:</strong></p><p><strong>Data Pipeline Development and Management:</strong></p><ul><li>Design, develop, and maintain robust and scalable data pipelines using Azure Data Factory and Azure Synapse.</li><li>Implement data ingestion, transformation, and loading (ETL/ELT) processes from various data sources into Snowflake.</li><li>Optimize data pipelines for performance, reliability, and cost-effectiveness.</li><li>Monitor and troubleshoot data pipeline issues.</li></ul><p><strong>Data Warehousing and Modeling:</strong></p><ul><li>Design and implement data models in Snowflake to support business intelligence and analytics needs.</li><li>Optimize Snowflake performance through query tuning and data partitioning.</li><li>Ensure data quality and consistency within the data warehouse.</li></ul><p><strong>Software Development and Automation:</strong></p><ul><li>Develop and maintain data processing applications using Python and Java.</li><li>Automate data engineering tasks using scripting and orchestration tools.</li><li>Implement and maintain CI/CD pipelines for data pipeline deployments.</li><li>Develop and maintain infrastructure as code.</li></ul><p><strong>Cloud Platform Management:</strong></p><ul><li>Manage and optimize Azure data services, including Data Factory, Synapse, and storage.</li><li>Implement and maintain data security and access controls in Azure and Snowflake.</li><li>Monitor cloud resource utilization and optimize costs.</li></ul><p><strong>Collaboration and Communication:</strong> </p><ul><li>Collaborate with data scientists, analysts, and other stakeholders to understand data requirements.</li><li>Communicate effectively with technical and non-technical audiences. </li><li>Participate in code reviews and knowledge sharing sessions.</li></ul>Sr Data Engineer<p>We are offering a contract employment opportunity for a Sr Data Engineer, EST time zone preferred. This role is deeply rooted in the data architecture and cloud data warehousing industry and involves extensive interaction with multiple departments such as IT, HR, Finance, Business Operations, Business Development, Sales, and Marketing. </p><p><br></p><p>Responsibilities:</p><p><br></p><p>• Design, develop, and implement enterprise data solutions using tools like Snowflake and DBT.</p><p>• Collaborate with cross-functional teams to ensure data solutions are comprehensive, accessible, accurate, and secure.</p><p>• Lead the establishment and management of the organization's first Enterprise Data Lake and Data warehouse.</p><p>• Manage key reference data for the enterprise, ensuring accuracy and accessibility.</p><p>• Design, develop, and maintain data integration solutions to facilitate seamless data flow between 20+ enterprise systems/applications.</p><p>• Ensure proficiency in stakeholder management, Azure database platforms, and programming languages related to data warehouse management.</p><p>• Handle the processing of customer credit applications efficiently and accurately.</p><p>• Maintain accurate customer credit records and monitor customer accounts.</p>Data Engineer<p>We are currently offering a permanent employment opportunity for a Data Engineer, onsite in Draper, Utah. The individual in this role will be primarily responsible for developing and optimizing data pipelines, designing ETL/ELT processes, and ensuring data accessibility. This position requires close collaboration with various teams to troubleshoot issues and refine data models. The work schedule for this IT position is Monday to Friday, from 9 AM to 5 PM.</p><p><br></p><p>Responsibilities:</p><p><br></p><p>• Develop and maintain data pipelines for seamless integration from Microsoft Azure, Salesforce, and external systems.</p><p>• Design, optimize, and implement ETL/ELT processes to keep data structured, normalized, and easily accessible.</p><p>• Collaborate with cross-functional teams to troubleshoot challenges and refine data models.</p><p>• Mentor other developers, sharing best practices and continuous improvement strategies for our data infrastructure.</p><p>• Support business intelligence and reporting needs by ensuring the data infrastructure is robust and efficient.</p><p>• Apply your expertise in Python for scripting and data processing tasks.</p><p>• Utilize ETL/ELT tools and frameworks to manage and transform data.</p><p>• Work with relational databases such as PostgreSQL, MySQL, SQL Server, and NoSQL databases like MongoDB.</p><p>• Implement Data Warehousing Solutions like Snowflake or Redshift and big data frameworks like Hadoop and Apache Spark.</p><p>• Employ containerization and orchestration tools like Kubernetes in your work.</p><p>• Implement Agile Development Framework in your work processes.</p><p>• Understand and apply data governance frameworks, like NIST CSF 2.0.</p><p>• Use GitHub or similar for CI/CD pipeline proficiency.</p><p>• Display excellent communication and collaboration skills to work effectively across teams.</p><p>• Demonstrate your ability to work with large datasets and complex business logic.</p><p>• Apply your strong problem-solving skills, with the ability to debug and optimize data processes.</p>Data EngineerWe are offering a long-term contract employment opportunity for a Data Engineer to join our team located in BATH TWP, Ohio. This role operates within the industry and demands the successful candidate to be skilled in developing SQL queries, managing HL7 files, and maintaining SSRS reports. <br><br>Responsibilities: <br>• Build and ensure the optimal function of our data platform to meet business and client needs.<br>• Create and manage SQL queries and Power BI reports for data analysis and visualization.<br>• Handle the manipulation and reporting of Health Level Seven (HL7) files.<br>• Maintain and update Legacy SSRS reports to ensure accuracy and efficiency.<br>• Adhere to and suggest improvements for best practice quality standards.<br>• Guarantee the accuracy and consistency of all documentation.<br>• Use a variety of skills including Apache Kafka, Apache Pig, Apache Spark, Cloud Technologies, and AWS Technologies.<br>• Implement analytics and algorithms for data processing.<br>• Develop APIs and utilize Fabric for optimal performance.<br>• Leverage SQL Server and Power BI for efficient data analysis and visualization.<br>• Utilize SSIS SSRS for streamlined data integration and reporting.Data EngineerWe are offering a contract for a permanent position as a Data Engineer in the logistics industry, located in Eagan, Minnesota. The selected candidate will play a crucial role in designing and expanding AI and ML capabilities, overseeing our data warehouse, and enhancing our analytical abilities.<br><br>Responsibilities:<br><br>• Design and expand AI and ML capabilities to enhance business functionality.<br>• Manage and oversee the company's data warehouse to ensure accurate and efficient data storage.<br>• Enhance analytical capabilities to provide insightful data-driven solutions.<br>• Migrate data workflows from Azure Synapse Analytics to Snowflake Analytics environment to improve data processing and analysis.<br>• Collaborate with business leaders to leverage data for informed decision-making.<br>• Drive platform and technology choices and own solution approaches and timelines to ensure optimal data management.<br>• Maintain technical documentation to provide clear and comprehensive data records.<br>• Lead BI solution design sessions, manage integrations, configurations, and API interfaces to improve data and AI accessibility.<br>• Implement data automation processes using Azure Data Factory and Azure Function API’s for data ingestion into Azure Data Lake.<br>• Work efficiently within an Agile environment to enhance team productivity.<br>• Utilize best practices in developing architectural solutions with a focus on enterprise usability.<br>• Project manage data-related projects, including migration to the Snowflake platform.Data EngineerWe are seeking a Data Engineer in New York, New York, 10006, United States. This role offers a contract to permanent employment opportunity, where you will be involved in data processing, modeling, governance, and visualization. You will work collaboratively with internal teams across various departments and be accountable for timely project delivery. <br><br>Responsibilities:<br>• Handle raw data by performing data cleansing and processing to produce large, structured datasets for analysis and data modeling.<br>• Understand the connectivity across multiple platforms like cloud-based storage environments, on-prem databases, etc., to extract and compile necessary data assets for advanced modeling projects.<br>• Create custom data integrations to support reporting, data visualization, and presentation building.<br>• Collaborate with internal teams, including BI, Buying, Planning, and Digital Investment to produce integrated client deliverables.<br>• Support R& D initiatives by setting up the appropriate infrastructure for algorithmic modeling, automating repetitive components of code, and establishing processes around self-service analytic capabilities.<br>• Utilize critical thinking skills to identify issues, solve problems, and improve processes.<br>• Maintain data documentation and definitions, and provide support to train business users on how best to use data.<br>• Take accountability for timely project delivery, producing analyses and content that is clean, organized, and easy to understand.<br>• Perform storytelling from data and analyses, and persuasively communicate insights to internal stakeholders and clients.<br>• Ensure satisfaction of internal stakeholders and clients.Senior Data EngineerWe are on the search for a Senior Data Engineer to join our team. In this role, you will be expected to leverage your analytical and problem-solving skills to manage and administrate our DBMS. Working in the IRVINGTON, New York, you will have the opportunity to design, develop, deploy, and optimize DB schemas, as well as liaise with developers to improve applications and establish best practices. <br><br>Responsibilities:<br><br>• Manage and administer the database management systems on a daily basis.<br>• Collaborate with application developers and business teams to assist with data-related technical issues and ensure successful design and delivery of data driven solutions.<br>• Design, develop, and deploy database schemas to meet application functionality and performance requirements.<br>• Create and implement SQL scripts and stored procedures.<br>• Develop and execute data reporting, data extract, data load and data validation processes and procedures.<br>• Implement processes and procedures for the development and release of products/projects that facilitate high quality and rapid deployment.<br>• Provide solutions to promote data integrity in enterprise systems, including data auditing, archive, backup, and restore solutions.<br>• Participate in code reviews to validate effectiveness and quality of code with internal team members as well as external vendor supported products and databases.<br>• Publish documentation and collaborative information using internal tools. <br>• Troubleshoot and resolve data related issues.Data Engineer<p>We are seeking a Data Engineer to be an integral part of our team in Dublin, Ohio. This role is pivotal in our IT Corporate Finance Reporting department, where you will provide technical and consultative support on complex matters. You will be involved in the design and development of systems based on user specifications, as well as providing assistance related to data and infrastructure needs. This role offers a short-term contract employment opportunity.</p><p><br></p><p>Responsibilities:</p><p><br></p><p>• Provide technical assistance to solve hardware or software problems.</p><p>• Analyze, design, and develop systems based on user specifications.</p><p>• Maintain knowledge of technical industry trends.</p><p>• Troubleshoot database issues related to connectivity and performance.</p><p>• Be involved in source control procedures using Git.</p><p>• Use Visual Studio for .NET and ETL SSIS development.</p><p>• Conduct Manual and Automated Testing.</p><p>• Work on enhancements or cross-impact projects.</p><p>• Utilise cloud technologies such as AWS, Azure, and Google Cloud Platform.</p><p>• Understand the processes and procedures within a corporate environment and work with different stakeholders.</p><p>• Have general database knowledge, including Microsoft SQL, Oracle, Snowflake, and Azure Data Lakes.</p><p>• Communicate effectively with team members and stakeholders.</p><p>• Support treasury applications, ensuring they run smoothly.</p>Azure Data Engineer<p>We are seeking a Data Engineer to join our team in the North Houston area. In this role, you will focus on creating, optimizing, and maintaining data pipeline architecture. This role is essential in supporting data initiatives across the company. </p><p><br></p><p>Responsibilities:</p><p><br></p><p>• Apply data architecture principles for modeling, stored procedures, replication, security, and compliance to meet both technical and business objectives.</p><p>• Construct and maintain scalable data pipelines using Microsoft Fabric components such as Lakehouse, Azure Data Factory (ADF), Data Warehouses, Notebooks, and Dataflows.</p><p>• Develop and implement solutions for data extraction, processing, and analysis of large volumes of structured and unstructured data.</p><p>• Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver high-quality data solutions.</p><p>• Optimize and troubleshoot data pipelines to ensure reliability, efficiency, and performance.</p><p>• Implement best practices for data governance and security to ensure data integrity and regulatory compliance.</p><p>• Oversee and maintain data infrastructure to ensure high availability and scalability.</p><p>• Stay updated with advancements in data engineering and Microsoft Fabric technologies.</p><p>• Develop technology architecture strategies informed by various business scenarios and motivations.</p><p>• Assess emerging technology trends and provide guidance on their potential impact on organizational opportunities and risks.</p><p>• Enhance the quality, consistency, accessibility, and security of data across the company continuously.</p><p>• Manage the Data Platform roadmap, ensuring future-proofing, capacity planning, and optimization.</p><p>• Work with IT Operations teams and BI vendors to resolve production issues.</p><p>• Manage stakeholder demands and prioritize reporting requirements and needs.</p><p>• Maintain relationships with key vendors to plan and adapt the Data Platform roadmap and leverage existing capabilities.</p><p>• Monitor Data Platform usage, deprecate unused reports and datasets, and drive continuous improvement.</p><p>• Work with users to address data issues, trace data lineage, and implement data cleansing and quality processes.</p><p>• Manage secure access to data sets for the business.</p><p>• Ensure effective collaboration with BI vendors, IT, and business teams.</p><p>• Manage project schedules, focus on critical path items, define and communicate key milestones, and coordinate with the broader team for tollgate reviews.</p><p>• Monitor and report on emerging technologies, seeking opportunities for continuous improvement.</p><p>• Ensure adherence to data visualization and data modeling processes, procedures, and standards.</p><p>• Manage data sets and work with ETL and Data Warehouse solutions.</p><p>• Launch and drive a data literacy program for business users.</p>Data EngineerWe are offering a contract to permanent employment opportunity for a Data Engineer in Philadelphia, Pennsylvania. As a Data Engineer, you will work in the healthcare industry, focusing on designing, developing, and delivering data solutions to support enterprise analytics. This role involves a strong emphasis on SQL and data engineering tools and methods, and the application of Python or Apache Spark for data transformations. <br><br>Responsibilities:<br>• Design, develop, and implement data engineering solutions and data warehousing on Databricks.<br>• Ensure accurate ingestion, integration, and curation of data to deliver valuable information to stakeholders.<br>• Develop efficient processes for data ingestion from diverse sources to Databricks on Azure.<br>• Provide guidance, coaching, and training to entry level and new team members on standards, processes, and documentation.<br>• Collaborate effectively with a large team of Data Engineers, Data Architects, and DevOps Engineers.<br>• Work towards developing expertise in Python or Apache Spark for data transformations.<br>• Ensure seamless integration of data across the enterprise and drive automation of repetitive tasks.<br>• If not already certified, aim to successfully obtain certification in Epic Clarity Data Model within the first six months of employment.<br>• Emphasis on providing research-ready datasets.Data Engineer-Advent/AxysWe are in search of a meticulous Data Engineer-Advent/Axys to be a part of our Investment Management team located in Evanston, Illinois. This role involves handling customer applications, maintaining customer records, and resolving customer inquiries with precision. The position also requires monitoring customer accounts and taking necessary actions. <br><br>Responsibilities:<br><br>• Efficiently process customer credit applications using Advent - Axys and Microsoft Excel.<br>• Utilize SQL - Structured Query Language to manage SQL databases.<br>• Maintain customer credit records with the highest level of accuracy.<br>• Resolve customer inquiries by demonstrating excellent listening skills.<br>• Monitor customer accounts and take appropriate actions when necessary.<br>• Display a high level of organization and attention to detail in all tasks.Senior Manager Data Engineer<p>We are looking for a highly skilled Data Engineering and Software Engineering professional to design, build, and optimize our Data Lake and Data Processing platform on AWS. This role requires deep expertise in data architecture, cloud computing, and software development, as well as the ability to define and implement strategies for deployment, testing, and production workflows.</p><p><br></p><p>Key Responsibilities:</p><ul><li>Design and develop a scalable Data Lake and data processing platform from the ground up on AWS.</li><li>Lead decision-making and provide guidance on code deployment, testing strategies, and production environment workflows.</li><li>Define the roadmap for Data Lake development, ensuring efficient data storage and processing.</li><li>Oversee S3 data storage, Delta.io for change data capture, and AWS data processing services.</li><li>Work with Python and PySpark to process large-scale data efficiently.</li><li>Implement and manage Lambda, Glue, Kafka, and Firehose for seamless data integration and processing.</li><li>Collaborate with stakeholders to align technical strategies with business objectives, while maintaining a hands-on engineering focus.</li><li>Drive innovation and cost optimization in data architecture and cloud infrastructure.</li><li>Provide expertise in data warehousing and transitioning into modern AWS-based data processing practices.</li></ul>Data Engineer<p>Are you a skilled Data Engineer with a passion for designing and implementing innovative data solutions in the cloud? Do you have hands-on expertise in Azure and experience with Medallion Architecture to optimize data processing pipelines? If so, we have an excellent opportunity for you to join a dynamic and forward-thinking team.</p><p>As an <strong>Azure Data Engineer</strong>, you’ll play a pivotal role in creating and maintaining scalable, efficient, and high-quality data workflows that enable actionable business insights. You will collaborate with cross-functional teams to design data lakehouses leveraging Medallion Architecture, transforming raw data into analysis-ready datasets.</p>Data EngineerWe are offering an exciting opportunity for a Data Engineer in Woodland Hills, California. The selected candidate will be an integral part of our team, contributing to the design and development of scalable data pipelines, collaborating with data scientists, and ensuring data governance. This role involves working within the industry to integrate data engineering solutions into our broader product architecture.<br><br>Responsibilities:<br>• Collaborate closely with data scientists to prepare datasets for model training, validation, and deployment<br>• Develop, design, and sustain scalable data pipelines to support dynamic pricing models<br>• Oversee and optimize ETL (Extract, Transform, Load) processes to assure data reliability and accuracy<br>• Contribute to best practices and document data engineering processes<br>• Engage with relevant stakeholders to comprehend data requirements and convert these into technical specifications<br>• Ensure adherence to data governance and compliance with appropriate data privacy and security regulations<br>• Integrate data engineering solutions into the broader product architecture in collaboration with the software development team<br>• Continuously monitor and troubleshoot data workflows to ensure reliable data integrationData Engineer<p>We are searching for a Data Engineer with skills in Fabric/Azure, Python, and cloud technologies to join our team. This role is in Chicago, Illinois and offers a short-term contract employment opportunity. As a Data Engineer, you will be tasked with developing data pipelines in Fabric, understanding the Fabric medallion architecture, and utilizing Python for various projects. </p><p><br></p><p>Responsibilities:</p><p><br></p><p>• Develop and manage data pipelines in Fabric, ensuring smooth deployment from development to production.</p><p>• Gain a comprehensive understanding of Fabric medallion architecture, including bronze, silver, and gold layers.</p><p>• Utilize your Python knowledge to enhance and optimize data processing tasks.</p><p>• Utilize your skills in Apache Kafka, Apache Pig, and Apache Spark for data processing and analytics.</p><p>• Use your knowledge of cloud technologies to manage and maintain the data infrastructure.</p><p>• Leverage your skills in data visualization and algorithm implementation to create insightful data reports and visualizations.</p><p>• Work with AWS technologies and API development to enhance data processing capabilities.</p><p>• Maintain excellent communication skills and be able to work remotely and independently under tight deadlines.</p><p>• Use your knowledge of Apache Hadoop and analytics to analyze large data sets and derive insights.</p><p>• Apply your skills in Microsoft Fabric and data pipelines to enhance data processing and storage capabilities.</p>Data EngineerWe are seeking a Data Engineer to join our team based in Bethesda, Maryland. As part of our Investment Management team, you will play a crucial role in designing and maintaining data pipelines in our Azure Data Lake, implementing data warehousing strategies, and collaborating with various teams to address data engineering needs.<br><br>Responsibilities:<br><br>• Design robust data pipelines within Azure Data Lake to support our investment management operations.<br>• Implement effective data warehousing strategies that ensure efficient storage and retrieval of data.<br>• Collaborate with Power BI developers to integrate data reporting seamlessly and effectively.<br>• Conduct data validation and audits to uphold the accuracy and quality of our data pipelines.<br>• Troubleshoot pipeline processes and optimize them for improved performance.<br>• Work cross-functionally with different teams to address and fulfill data engineering needs with a focus on scalability and reliability.<br>• Utilize Apache Kafka, Apache Pig, Apache Spark, and other cloud technologies for efficient data visualization and algorithm implementation.<br>• Develop APIs and use AWS technologies to ensure seamless data flow and analytics.<br>• Leverage Apache Hadoop for effective data management and analytics.