Data Scientist- VLM (Vision Language Model)
Waukesha, WI Jobs
About the job you're considering
We are seeking a highly skilled and detail-oriented Vision-Language Models (VLM) Data Scientist/ Vision Data Analyst to join our team. The ideal candidate will have a strong background in computer vision, natural language processing, data analysis, and machine learning.This role involves developing and deploying multimodal AI solutions that integrate vision and language capabilities, analyzing visual data to extract meaningful insights, and collaborating with cross-functional teams to improve our products and services.
Your role
· VLM Development & Deployment: Design, train, and deploy efficient Vision-Language Models (e.g., VILA) for multimodal applications. Explore cost-effective methods such as knowledge distillation, modal-adaptive pruning, and LoRA fine-tuning to optimize training and inference.
· Multimodal AI Solutions: Develop solutions that integrate vision and language capabilities for applications like image-text matching, visual question answering (VQA), and document data extraction. Leverage interleaved image-text datasets and advanced techniques (e.g., cross-attention layers) to enhance model performance.
· Healthcare Domain Expertise: Apply VLMs to healthcare-specific use cases such as medical imaging analysis, position detection, motion detection, and measurements. Ensure compliance with healthcare standards while handling sensitive data.
· Efficiency Optimization: Evaluate trade-offs between model size, performance, and cost using techniques like elastic visual encoders or lightweight architectures. Benchmark different VLMs (e.g., GPT-4V, Claude 3.5) for accuracy, speed, and cost-effectiveness on specific tasks.
· Data Analysis: Analyze large sets of visual data to identify patterns, trends, and anomalies.
Algorithm Development: Develop and implement computer vision algorithms to process and interpret visual data.
· Machine Learning: Apply machine learning techniques to improve the accuracy and efficiency of vision-based systems.
· Reporting: Create detailed reports and visualizations to communicate findings to both technical and non-technical audiences.
Your skills and experience
· Education: Master's or Ph.D. in Computer Science, Data Science, Machine Learning, Electrical Engineering, or a related field.
· Experience: 3+ years of experience in machine learning or data science roles with a focus on vision-language models and computer vision. Proven expertise in deploying production-grade multimodal AI solutions.
· Technical Skills: Proficiency in Python and ML frameworks (e.g., PyTorch, TensorFlow). Hands-on experience with VLMs such as VILA, or VSS. Strong understanding of image processing techniques and tools.
· Analytical Skills: Excellent problem-solving skills and the ability to analyze complex data sets.
Communication: Strong written and verbal communication skills. Ability to present complex information clearly and concisely.
· Teamwork: Ability to work effectively in a collaborative team environment.
· Experience with cloud computing platforms such as AWS or Azure.
· Familiarity with data visualization tools like Tableau or Power BI. Knowledge of statistical analysis and data mining techniques.
Data Scientist
Richmond, VA Jobs
Data Scientist.
Duration: 12 months
Short Description:
Data Scientist (Skill level 4) for The Client Professional Services Procurement Office(PSPO); Procurement knowledge / experience strongly preferred.
Complete Description:
This Data Scientist position with the Professional Services Procurement Office will be an important member of the Client PSPO team. In this role, the Data Scientist collects and analyzes statistics and information from multiple sources to spot trends and to gain maximum insight that can give the company a competitive advantage and communicates informed conclusions and recommendations across an organization's leadership structure. Strategizes and identifies unique opportunities to locate and collect new data, explores and mines data from many angles, and determines what it means. Communicates data findings to both business and IT leaders to influence how an organization approaches and meets business challenges of an evolving customer base and changing marketplace, using strong business acumen. Finds and recommends new uses for existing data sources; designs, modifies, and builds new data processes; and builds large, complex data sets.
In addition to standard knowledge, skills and abilities for a Data Scientist, the successful candidate for this role will have experience and knowledge of procurement, preferably professional services procurement and the data related to this business need as well as procurement related reporting systems and technology.
Skills Specific to this Position:
Provide expertise to agency on data-driven decision making and strategy.
Provide expertise in specific areas, such as machine learning, predictive analytics using data and systems related to public procurement.
Leading research projects, for advanced date mining, developing algorithms, or building data pipelines related to public procurement.
Required Skills:
7 Years Statistics and probability
7 years Database management
7 Years Machine learning and deep learning
7 Years Data Visualization
7 Years Cloud Computing
7 Years Interpersonal Skills
7 Years Predictive Analytics
7 Years Data Cleansing and Transformation
7 Years Cross Functional Collaboration
7 Years Business Intelligence and Reporting
7 Years Statistical Modeling
7 Years Experience utilizing these skills in public sector procurement environment
Data Engineer III (Glider assessment required).
Charlotte, NC Jobs
Immediate need for a talented Data Engineer III (Glider assessment required). This is a 12+months contract opportunity with long-term potential and is located in Charlotte, NC (Hybrid). Please review the job description below and contact me ASAP if you are interested.
Job ID: 25-72960
Pay Range: $60 - $70/hour. Employee benefits include, but are not limited to, health insurance (medical, dental, vision), 401(k) plan, and paid sick leave (depending on work location).
Key Responsibilities:
This position is for an experienced Data Engineer that will work on the EcoSys Project and be engaged in integrating various on-prem financial and project management systems with an external vendor solution.
The work requires not only feeding the vendor solution but reworking existing solutions in the same subject area.
Key Requirements and Technology Experience:
Key Skills; AWS Services, Data Integration Experience, Data Design Experience, REST APIs , SQL, designing, building, and maintaining databases, S3, EMR, Glue Jobs, Lambda, Aurora, CloudTrail.
Applicants should be self-motivated individuals with excellent communication skills who possess the desired qualifications in this requisition and who work well in a collaborative environment.
Five (5) or more years of AWS experience using one or more of the associated services - S3, EMR, Glue Jobs, Lambda, Aurora, CloudTrail, SNS, SQS, CloudWatch
Experience with Databases such as Redshift, PostgreSQL, SQL Server, Oracle
Experience using REST APIs using Glue and/or Lambdas
Eight (8) or more years designing, building, and maintaining enterprise-scale databases
Eight (8) or more years of experience in data warehouse architecture
Eight (8) or more years SQL experience
Four (4) or more years SQL Server Analysis Services (SSAS)
Excellent customer service, requirements gathering, and data analysis skills
Excellent analytical abilities and documentation skills
Experience with Kafka/Messaging preferably Confluent Kafka
Competence in developing and refining data pipelines within AWS
Proficient in handling both real-time and batch data processing tasks
Proficient in Advanced features of Microsoft Excel
Python, NodeJS
Workflow Automation
Experience transitioning on premise platforms into cloud-based platforms
Power BI and SSRS
Experience using AWS monitoring services like CloudWatch, CloudTrail ad CloudWatch events
Basic nderstanding of networking DNS, TCP/IP and VPN
Experience with AWS workflow orchestration tool like Airflow or Step Functions
Experience with Secrets Management Platform like Vault and AWS Secrets manager
Experience with Event Driven Architecture
Our client is a leading Utility Industry, and we are currently interviewing to fill this and other similar contract positions. If you are interested in this position, please apply online for immediate consideration.
Pyramid Consulting, Inc. provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.
By applying to our jobs you agree to receive calls, AI-generated calls, text messages, or emails from Pyramid Consulting, Inc. and its affiliates, and contracted partners. Frequency varies for text messages. Message and data rates may apply. Carriers are not liable for delayed or undelivered messages. You can reply STOP to cancel and HELP for help. You can access our privacy policy here.
Data Scientist
Stamford, CT Jobs
Onsite
We are seeking a skilled and experienced Data Scientist with ~5 years of industry experience to develop reporting/tools as well as actively participation in the design and measurement of campaigns. End-product will be spreadsheets and/or dashboards that can take inputs (SKU participation, offer value, profitability) and estimate resulting volume. Secondary functions include assisting in media and consumption forecasting.
The role is on-location in Stamford, CT, with the expectation of 4 days a week in the office to start, tapering to 3 days once
EXPERIENCE
Managing business processes through reporting, analysis
~5 Years in analytics/analyst roles
A/B Testing experience
CPG Experience a plus
Regression modeling and similar experience a plus
SKILLS
People skills - comfortable interacting with brand teams, retailer teams, and other stakeholders
Coordinate workstreams, and manage communication between internal business units and external agencies
Expert Excel skills
PowerBI development
Hands-on experience in building ETL/ELT pipelines for data ingestion
Expertise in Python or R for data manipulation, analysis, and model development
Sr Data Scientist
Richardson, TX Jobs
Title: Sr Data Scientist
Telecom domain must… Need have solid exp on RAG systems, transformers, and NLP models
We're looking for a highly skilled and experienced Data Scientist to help lead this transformation. If you're passionate about turning complex data into actionable insights, advancing the frontiers of anomaly detection, transformers, and Retrieval-Augmented Generation (RAG), and pushing the boundaries of what AI can do, this is your opportunity. Telecom experience is a must, but what matters most is your curiosity, creativity, and deep expertise in machine learning.
Job Responsibilities
Develop and optimize RAG (Retrieval-Augmented Generation) based systems to improve information retrieval and decision-making processes.
Apply transformers and deep learning models to extract insights from large and complex datasets.
Implement anomaly detection algorithms to identify unusual patterns and potential issues in telecom and other datasets.
Design and deploy scalable AI/ML solutions that address real-world business challenges.
Collaborate with cross-functional teams, including data engineers, software developers, and domain experts, to integrate ML solutions into production environments.
Stay up to date with the latest advancements in AI/ML and apply them to enhance model performance.
Communicate findings and recommendations effectively to both technical and non-technical stakeholders.
Job Requirements
Bachelor's degree in Computer Science, Data Science, AI/ML, or a related technical field from an accredited university, or equivalent work experience.
Minimum 8 years of relevant work experience.
Strong experience in machine learning, deep learning, and AI frameworks (e.g., TensorFlow, PyTorch, Scikit-Learn).
Hands-on experience with RAG systems, transformers, and NLP models.
Expertise in anomaly detection techniques, including statistical and ML-based approaches.
Strong background in data preprocessing, feature engineering, and model evaluation.
Training/Pretraining/Fine-tuning LLMs and ML experience with large datasets.
Excellent problem-solving and analytical skills.
Networking technology background.
Preferred Qualifications:
Advanced degree (MS/PhD) in Computer Science, Data Science, AI/ML, or a related field.
Familiarity with graph databases and knowledge graphs for information retrieval.
Contributions to research publications, open-source projects, or AI/ML communities.
Understanding of AI governance and implementation.
Raj Vemula
Director-Resource Development
*************************
******************
Python and lead Data Engineer
Woodbridge, NJ Jobs
Job Description
Software Engineer 3 Python and Data lead
Only one round- F2F Required
W2 Candidates (USC,GC,GC EAD,H4 EAD)- with minimum validity of 12 months
Skills:Python, bash , various ETL frame, dremio , Kafka Api , Apigee experience actually going work with data lead, implement the data from various system
7+ Proficiency in scripting languages such as Python, Bash, Shell
5+ years of experience working in Agile environment using Scrum/Kanban
5+ years of experience working in Jira, Confluence
Python Data Engineer
Charlotte, NC Jobs
Job Description
Title: Python Data Engineer Look for Local / nearby Candidates (Only w2 with 12 months of visa validity) Duration: 12+ Months Contract Visa: USC,GC,GC EAD,H4 EAD(Only W2)
Skills:
Requires Pyspark Python S3 Dremio ETL skills.
Some financial exp would be beneficial.
Profit View modernization effort
Resource will be responsible for end-to-end development covering portions of Financial Attribution
SCD Booking and Referring Agreements Data Aggregations and SOR Onboarding.
In this contingent resource assignment you may:
Consult on complex initiatives with broad impact and large-scale planning for Software Engineering.
Review and analyze complex multi-faceted larger scale or longer-term Software Engineering challenges that require in-depth evaluation of multiple factors including intangibles or unprecedented factors.
Contribute to the resolution of complex and multi-faceted situations requiring solid understanding of the function policies procedures and compliance requirements that meet deliverables.
Strategically collaborate and consult with client personnel.
Required Qualifications: 5 years of Software Engineering experience or equivalent demonstrated through one or a combination of the following: work or consulting experience training military experience education.
Data Engineer
Newport Beach, CA Jobs
Job Title: Data Engineer
Reports to: Sr. Manager, Software Engineering in Technology
Job Type: Full-time
Salary Range: $120K-$140K
About Our Organization: RIS Rx (pronounced “RISE”) is a healthcare technology organization with a strong imprint in the patient access and affordability space. RIS Rx has quickly become an industry leader in delivering impactful solutions to stakeholders across the healthcare continuum. RIS Rx is proud to offer an immersive service portfolio to help address common access barriers. We don't believe in a “one size fits all” approach to our service offerings. Our philosophy is to bring forward innovation, value and service to everything that we do. This approach has allowed us to serve countless patients to help produce better treatment outcomes and improved quality of life. Here at RIS Rx, we invite our partners and colleagues to “Rise Up” with us to bring accessible healthcare and solutions for all.
Summary: We are seeking a highly skilled Data Engineer to join our team and play a crucial role in building and maintaining scalable, high-performance data pipelines and data lake architectures. You will be working with large data sets across multiple products, ensuring real-time data processing, data quality, and governance that support machine learning and model training efforts. The ideal candidate has deep expertise in SQL, PostgreSQL, AWS, data lakes, and big data tools like Spark, along with strong proficiency in Python and Golang.
Duties and Responsibilities include but are/ not limited to the following:
Design, build, and maintain scalable ETL pipelines for processing structured and unstructured data.
Develop and optimize data lakes using AWS S3, AWS Lake Formation, and Glue to enable efficient data storage and retrieval.
Develop real-time data processing solutions to handle streaming data efficiently.
Automate and monitor data workflows to ensure system reliability and performance.
Optimize and manage PostgreSQL databases for performance and scalability.
Utilize AWS services (e.g., S3, Glue, Lambda, Kinesis) for cloud-based data processing.
Work with Apache Spark or similar big data processing frameworks.
Ensure data quality, governance, and security best practices are implemented for data lakes and pipelines.
Collaborate with Engineering team to support model training and deployment.
Write efficient, well-structured SQL queries for data analysis and transformation.
Develop and maintain data infrastructure using Python, Golang, Terraform, and Pulumi.
Qualifications/Skills:
To perform this job successfully, an individual must be able to perform each essential duty satisfactorily. The requirements listed below are representative of the knowledge, skill, and/or ability required. Reasonable accommodations may be made to enable individuals with disabilities to perform the essential functions.
5+ years of experience in data engineering or a similar role.
Strong expertise in SQL and database design, specifically in Postgres.
Hands-on experience with AWS big data tools (e.g., S3, Glue, Lake Formation, Lambda, Kinesis).
Proficiency in big data processing frameworks such as Apache Spark.
Strong programming skills in Python/Golang and Terraform/Pulumi.
Experience with real-time data processing using tools like Kafka or Kinesis.
Implemented data governance, data quality, and security best practices.
Experience working with machine learning pipelines and supporting model training.
Experience with Scrum and Agile processes.
Preferred experience with Power BI.
Effective communication and problem-solving skills and the ability to work in a fast-paced environment.
Data Architect / Data Analyst ( experience in Hadoop required)
Irving, TX Jobs
TITLE: Data Architect / Data Analyst
POSITION TYPE: Full Time (W2)
ABOUT WorldLink:
WorldLink is a rapidly growing information technology company at the forefront of the tech transformation. From custom software development to cloud hosting, from big data to cognitive computing, we help companies harness and leverage today's most cutting-edge digital technologies to create value and grow.
Collaborative. Respectful. Work hard Play hard. A place to dream and do. These are just a few words that describe what life is like at WorldLink. We embrace a culture of experimentation and constantly strive for improvement and learning.
We take pride in our employees and their future with continued growth and career advancement. We put TEAM first. We are a competitive group that like to win. We're grounded by humility and driven by ambition. We're passionate, and we love tough problems and new challenges. You don't hear a lot of "I don't know how" or "I can't" at WorldLink. If you are passionate about what you do and having fun while doing it; tired of rigid and strict work environments and would like to work in a non-bureaucratic startup cultural environment, WorldLink may be the place for you.
For more information about our craft, visit *************************
WHO we're looking for:
We are looking for a Data Architect / Data Analyst to join our team.
Role and Responsibilities:
Lead and guide data architecture design and principles, data integrity, reference data, data model version, and metadata standards for data platform.
Collaborate with business and development teams to address data issues and data feed changes.
Collaborate with architecture and business users to review architecture design, evaluate technology tools, and technical solutions to influence business decisions.
Lead and execute data management strategy, define data quality measures, develop automated data validation control process using programing languages and tools, and dashboard monitoring process with automated alerts.
Identify critical data elements, create data catalog and data lineage for reporting and downstream data feeds.
Evaluate and review overall data flow adhering with data architecture standards, review with stakeholders, and enhance data governance oversight with change management process.
Assess complex data flows and identify opportunities to simplify end-to-end data processes.
Perform data and business process analysis to identify data issues and root causes and implement resolutions.
Contribute design and development of technical and business solutions for data and reporting.
Engage with global engineering teams and business partners to create business and functional requirements and solutions.
Led project plans with cross program interdependencies for end-to-end implementation.
Develop scalable and resilient data pipelines with real-time and batch processes into data warehouse.
Support reporting application and downstream changes and testing.
Required Experience and Education:
7-8 years of experience in data architecture and data analysis.
Bachelor's degree in computer science, engineering, or related field required.
Expertise in scalable data and event-driven enterprise architecture design, data management and data strategy.
Experience using Hadoop, Hive, and SQL.
Visualization knowledge - Tableau and Power BI.
Strong experience in leading the implementation of enterprise data architecture and data engineering utilizing best engineering practices.
Strong experience in design data management solutions and implementing including data control, data lineage, data catalogue, and security access.
Expertise in modern data architecture and technology.
Experience in data modeling for a large data warehouse, operational and analytical data stores.
Necessary Skills and Attributes:
Self-motivated individual with the ability to thrive in a team-based or independent environment.
Detail-oriented with strong organization skills.
Ability to work in a fast-paced environment.
Limited supervision and the exercise of discretion.
Excellent verbal and written communication skills, enabling effective collaboration with cross-functional teams, stakeholders, and management.
Proven ability to effectively use complex analytical, interpretive and problem-solving techniques and execute plans.
Passion and commitment for process and design simplification.
Data Architect (Manufacturing)
Dallas, TX Jobs
We are looking for an experienced SCADA & Data Architect to design and implement industrial data architectures that integrate SCADA (Ignition), IoT, Unified Namespace (UNS), and cloud data platforms. The ideal candidate will have hands-on experience with Ignition SCADA, HiveMQ (UNS), Azure Data Factory, and Snowflake, ensuring seamless data flow, scalability, and real-time analytics across industrial systems.
Key Responsibilities:
Design, develop, and optimize SCADA (Ignition) and IoT data architectures for industrial automation.
Implement Unified Namespace (UNS) using HiveMQ to enable real-time data streaming and interoperability.
Integrate SCADA and IoT data pipelines with Azure Data Factory and Snowflake for efficient storage, processing, and analytics.
Develop robust data ingestion, transformation, and processing workflows for industrial systems.
Ensure high availability, security, and scalability of industrial data platforms.
Work with MQTT, OPC UA, and other industrial communication protocols for seamless data exchange.
Collaborate with cross-functional teams to optimize industrial data workflows and analytics.
Provide technical leadership and best practices for SCADA, IoT, and cloud data integration.
Essential Skills & Qualifications:
Strong experience with Ignition SCADA for industrial automation and control systems.
Hands-on expertise with HiveMQ and Unified Namespace (UNS) for IoT and data streaming.
Experience in designing and implementing Azure Data Factory workflows.
Proficiency in Snowflake for industrial data storage, ETL, and analytics.
Knowledge of MQTT, OPC UA, and industrial IoT protocols.
Experience with real-time data processing and event-driven architectures.
Strong understanding of industrial data modeling, pipelines, and cloud integration.
Proficiency in Python, SQL, or scripting languages for data processing.
Preferred Qualifications:
Experience with edge computing and IoT gateways.
Knowledge of cybersecurity best practices for industrial data architectures.
Familiarity with DevOps/MLOps practices for data pipeline automation.
Azure Data engineer
Dallas, TX Jobs
Job Title: Azure Data engineer
Duration : Long term
Roles and Responsibilities.
Develop and manage Snowflake stored procedures, UDFs, and performance tuning techniques.
Implement access controls, row-level security, and data masking policies for secure data handling.
Integrate Snowflake with Azure Data Factory, external APIs, and data lakes for hybrid cloud solutions.
Good understanding on Snowflake RBAC and Roles and User Management.
Design and orchestrate scalable data pipelines using ADF for batch and near real-time data ingestion.
Develop ADF pipelines for data integration, transformation, and movement across multiple sources.
Implement parameterized pipelines, Linked Services, Datasets, and Triggers to support modular data workflows.
Monitor, troubleshoot, and optimize ADF pipeline performance and cost-efficiency.
Integrate ADF with Synapse, Snowflake, Azure Key Vault, and Logic Apps as part of end-to-end solutions.
Develop and optimize Synapse SQL scripts, stored procedures, and data flows for analytics workloads.
Implement Synapse Pipelines and Spark Notebooks for advanced transformation and machine learning use cases.
Perform data warehousing activities including partitioning, indexing, and materialized views.
Integrate Synapse with Power BI, Azure ML full-stack data insights.
Ensure efficient utilization of Dedicated SQL Pools and Serverless resources.
Synapse Administration ,RBAC understanding and User management from Azure Entra.
Set up and maintain CI/CD pipelines for ADF.
Implement version control, code branching strategies, and peer code reviews for collaborative development.
Configure release pipelines to handle environment-specific configurations using variable groups and templates.
Monitor pipeline runs, manage build/release failures, and enforce security and compliance during deployments.
Senior Data Engineer with Data Quality
New York, NY Jobs
Title: Senior Data Quality Specialist
Duration: 6-12+ months
Note:
Senior Data Quality Specialist 1 year - 3 Days onsite in New York • 8+ years of experience in data engineering, specifically with Azure Data Factory and Databricks. • Strong background in database and application server administration, with hands-on experience in managing SQL and NoSQL databases. • Demonstrated experience in data quality management, including data profiling, cleansing, and validation techniques. • Collibra/Owl DQ experience • API development with Spring boot experience You cannot use third party candidates. Candidates must be direct to you and not a subcontractor candidate.
Contract Description:
We are seeing a Senior Data Engineer contractor with expertise in data quality and database management. The ideal candidate will have a strong background in data engineering, particularly with Azure technologies and will be responsible for ensuring the integrity and quality of our data assets.
Design, develop, and maintain data pipelines using Azure Data Factory and Databricks to support data integration and transformation processes.
Administer and manage database and application servers, ensuring optimal performance and security.
Implement and monitor data quality frameworks to ensure accuracy, completeness, and reliability of data across the organization.
Collaborate with cross-functional teams to understand data requirements and provide data solutions that meet business needs.
Support the integration of Collibra for data governance and stewardship, ensuring compliance with data management policies
Qualifications:
8+ years of experience in data engineering, specifically with Azure Data Factory and Databricks.
Strong background in database and application server administration, with hands-on experience in managing SQL and NoSQL databases.
Demonstrated experience in data quality management, including data profiling, cleansing, and validation techniques.
Collibra/Owl DQ experience
API development with Spring boot experience
This role requires being on site in New York 3 days a week.
Senior Data Engineer
Santa Monica, CA Jobs
Full-time or W2 hourly, directly with Tavant ONLY
Should be in Santa Monica or be willing to relocate.
The interview will involve Hackerrank coding
● 12+ years of relevant data engineering experience.
● Deep coding knowledge of the Python/Snowflake/Databricks (all 3 are mandatory)
● Deep coding knowledge of databases (SQL and no-SQL)
● Experience in building large datasets and scalable batch pipelines.
● Experience with large-scale distributed data processing systems, their benefits and limitations.
Preferred Qualifications
● Experience with Airflow, or any other data pipeline orchestration tools.
● Experience designing and implementing cross-cloud data pipelines in AWS, Snowflake, Databricks delta lake files, etc.
Required Education
● Bachelor's degree in Computer Science, Software Engineering, or a related field.
Data Engineer
Redmond, WA Jobs
Role: Data Engineer
Duration: Long-term project.
Key Skills
Apache Spark (PySpark or Scala): Writing scalable distributed rule engines.
Apache Hive: Querying partitioned datasets, schema handling.
Amazon EMR: Cluster management, tuning jobs for large datasets (20+ TB).
AWS S3: Data partitioning, object lifecycle, access patterns.
Glue Data Catalog: Managing Hive table metadata.
Data Quality Frameworks: (Great Expectations, Deequ, or custom rules).
CI/CD: Code deployment pipelines (Git, Jenkins, Airflow, etc
AWS Snowflake and DBT Data Engineer
Houston, TX Jobs
Role: AWS Snowflake and DBT Data Engineer
Infosys is seeking an AWS Snowflake and DBT Data Engineer. In this role, you will enable digital transformation for our clients in a global delivery model, research on technologies independently, recommend appropriate solutions and contribute to technology-specific best practices and standards. You will be responsible to interface with key stakeholders and apply your technical proficiency across different stages of the Software Development Life Cycle. You will be part of a learning culture, where teamwork and collaboration are encouraged, excellence is rewarded, and diversity is respected and valued.
Required Qualifications:
Candidate must be located within commuting distance of Houston, TX or be willing to relocate to the area. This position may require travel in the US
Bachelor's degree or foreign equivalent required from an accredited institution. Will also consider three years of progressive experience in the specialty in lieu of every year of education.
All applicants authorized to work in the United States are encouraged to apply
At least 4 years of Information Technology experience
At least 2 years of AWS Glue experience
At least 2 years of Snowflake experience
Hands on experience with developing complex transformations using Data Build tool (DBT)
Experience in end-to-end implementation using technologies such as AWS and Snowflake,
Strong knowledge and hands-on experience in SQL, Unix shell scripting, and Python
Preferred Qualifications:
Experience in Services and Utilities domain
Strong knowledge Data structures, Data Engineering concepts, Algorithms, Collections, Multi-threading and memory management and concurrency
Experience in large scale cloud data migrations using Snowflake, Python, Spark, SQL
Sound knowledge of software engineering design patterns and practices
Good understanding of Agile software development frameworks
Strong communication and Analytical skills
Ability to work in teams in a diverse, multi-stakeholder environment comprising of Business and Technology teams
Lead Data Engineer
Atlanta, GA Jobs
Job Title: : Lead Data Engineer
Duration : Fulltime
We are seeking a Lead Data Engineer with expertise in data modeling, Data Vault architecture, and cloud data warehouse (CDW) architecture utilizing the Kimball model. The ideal candidate will have hands-on experience with data fabric technologies (e.g., Databricks), Spark, and Python on Azure Cloud. This role involves designing, implementing, and optimizing data pipelines and data platforms to support business intelligence and advanced analytics.
Key Responsibilities:
Design and implement scalable and efficient data models, including Data Vault and Kimball dimensional modeling for cloud-based data warehouses.
Architect and manage Azure Cloud-based data infrastructure, ensuring scalability, security, and performance.
Develop and optimize data pipelines using Databricks, Spark, and other data fabric tools to process large volumes of data.
Collaborate with data scientists, analysts, and business stakeholders to understand data needs and deliver high-quality solutions.
Implement best practices for data engineering, including data quality, governance, and lifecycle management.
Build and maintain infrastructure-as-code using Terraform or other DevOps tools.
Create reusable data engineering components for ingestion, transformation, and orchestration.
Troubleshoot, optimize, and maintain existing data systems to meet evolving business needs.
Required Skills and Experience:
10+ years of experience in Data Engineering or a similar role, with a proven track record of leading data projects.
Expertise in Data Vault 2.0 architecture and Kimball modeling for data warehouse design.
Proficiency with Azure Data Services (e.g., Azure Data Factory, Azure Synapse, Azure Databricks).
Hands-on experience with Databricks and Spark for distributed data processing.
Strong programming skills in Python for ETL/ELT development and automation.
Deep understanding of CDW architecture and data fabric frameworks.
Experience with CI/CD pipelines and infrastructure-as-code tools like Terraform.
Solid knowledge of data governance, security, and compliance best practices.
Educational Qualifications:
Bachelor's degree in Computer Science, Data Engineering, or related fields. Master's degree preferred.
Certifications in Data Engineering, Azure Cloud, or Databricks are highly desirable.
Data Engineer
Cleveland, OH Jobs
Full time
8+ Years experience working in Data Engineering and Data Analysis .
Hands on Experience in Hadoop Stack of Technologies ( Hadoop ,Spark, HBase, Hive , Pig , Sqoop, Scala ,Flume, HDFS , Map Reduce).
Hands on experience with Python & Kafka .
Good understanding of Database concepts , Data Design , Data Modeling and ETL.
Hands on in analyzing, designing, and coding ETL programs which involves Data pre-processing , Data Extraction , Data Ingestion , Data Quality ,Data Normalization & Data Loading.
Working experience in delivering projects in Agile Methodology and hands on in Jira.
Experience in Client Facing Roles with good communication & thought leadership skills to co-ordinate deliverables across the SDLC .
Good understanding of Machine Learning Models and Artificial Intelligence preferred.
Good understanding of Data Components , Data Processing & Data Analytics on AWS is good to have .
Experience with data modeling tools like Erwin is good to have.
Preferred Location : Cleveland or Pittsburgh.
Master's/bachelor's in computer science or equivalent fields
Salesforce Data Cloud Architect
Hartford, CT Jobs
Job Title : Salesforce Data Cloud Architecture
Salary: Market
Need immediate joiner
We are seeking an experienced Salesforce Data Cloud professional to join our team The ideal candidate will have a deep understanding of Salesforce Marketing Cloud SFMC or Pardot or Marketing Cloud Personalization with a focus on data management customer segmentation and analytics within the Salesforce Data Cloud environment This role will involve designing and implementing data solutions that drive customer engagement optimize marketing efforts and enhance data driven decision making
Key Responsibilities
1 Salesforce Data Cloud Implementation Lead the deployment and configuration of Salesforce Data Cloud ensuring seamless integration with existing Salesforce Marketing Cloud Pardot or any other platforms
2 Data Management Integration Oversee the collection cleansing and management of customer data from various sources ensuring data accuracy consistency and accessibility within Salesforce Data Cloud
3 Customer Segmentation Personalization Develop advanced customer segmentation strategies using Salesforce Data Cloud to enable personalized marketing campaigns and customer journeys
4 Analytics Reporting Create comprehensive dashboards and reports within Salesforce Data Cloud to track key performance metrics customer behaviours and campaign effectiveness
5 Data Governance Compliance Ensure adherence to data governance policies including data privacy regulations like GDPR and CCPA within Salesforce Data Cloud and related marketing platforms
6 Cross Functional Collaboration Work closely with marketing sales and IT teams to align data strategies with business objectives and drive marketing automation initiatives
7 Continuous Improvement Stay up to date with Salesforce product updates best practices and industry trends to continuously enhance the organizations use of Salesforce Data Cloud
8 Training Support Provide training and support to marketing and sales teams on the use of Salesforce Data Cloud ensuring they can effectively leverage data for their initiatives
Required Skills Qualifications
1 Hands on experience in Salesforce Marketing Cloud SFMC or Pardot or MCP or any other Leading Marketing Automation Platform with a strong focus on data management and customer analytics
2 Salesforce Data Cloud Expertise Proven experience with Salesforce Data Cloud Customer Data Platform including its implementation configuration and optimization
3 Technical Skills Proficiency in SQL data modelling and data integration techniques Experience with APIs and ETL processes is a plus
4 Analytical Mindset Strong analytical skills with the ability to interpret complex data sets and translate them into actionable marketing strategies
5 Communication Excellent verbal and written communication skills with the ability to convey technical concepts to nontechnical stakeholders
Nice to Have
1 Experience in Snowflake and MDM system
2 Experience in Enterprise Integration Patterns with API ETL logic implementation
3 Certification Salesforce Marketing Cloud Consultant Salesforce Pardot Consultant or Salesforce Data Cloud certifications are highly desirable
Skills
Mandatory Skills : Salesforce Automotive Cloud,Salesforce Communication Cloud,Salesforce Consumer Goods Cloud,Salesforce Education Cloud,Salesforce Energy and Utilities Cloud,Salesforce Financial Services Cloud,Salesforce Health Cloud,Salesforce Manufacturing Cloud,Salesforce Media Cloud,Salesforce Revenue Cloud,Salesforce Trade Promotion Management,Salesforce Data Architecture,Estimation,Integration Architecture,Business Process Design, Mapping,Roadmap/ Strategy,Salesforce Data Cloud,Data Architecture
Good to Have Skills : Salesforce Data Architecture
Data Engineer - ONLY W2
Dallas, TX Jobs
Data Engineer (Informatica)
Long Term Project
MUST HAVE: Informatica, python, SQL, Snaplogic, Knowledge of mainframe environment, using COBOL, DB2, CICS, JCL, and VSAM files
The Expertise and Skills You Bring
Bachelor's degree in a technology-related field
Extensive experience in ETL technologies (Informatica)
Experience with Java, Python, Relational Databases, SQL, Angular, and Data Analysis Tools (Tableau)
Experience in agile methodologies (Kanban and SCRUM)
Experience with Snap Logic, Kafka, Azure, AWS a plus
Deep knowledge of scrum tools - Jira
Domain knowledge of Retirement Services is a plus.
Ability to take care of ambiguity and work in fast paced environment.
Deep experience supporting mission critical applications.
Excellent communication skills, both through written and verbal channels
Ability to understand and adapt to changing business priorities and technology advancements.
Strategic thinking and critical problem-solving skills
Solid understanding of crafting and developing highly scalable, distributed solutions and integration with other middleware and UI applications/solutions
You will stand out if you have:
Extensive experience in ETL technologies (Informatica)
Solid Understanding of Unix/Linux operating systems and shell scripting
Knowledge of mainframe environment, using COBOL, DB2, CICS, JCL, and VSAM files
Data Engineer
Redmond, WA Jobs
Full Time with People Tech Group
Role: Data Engineer
Responsibilities
Qualifications:
Cloud Computing Experience, Azure or AWS preferred
Data Warehousing Experience, Snowflake preferred
Python Programming Experience, 5+ years preferred
Independent Worker who can critique and contribute to architectural design proposals