Data Architect
HAVI is a global, privately owned company focused on innovating, optimizing and managing the supply chains of leading brands. Offering services in marketing analytics, packaging, supply chain management and logistics, HAVI partners with companies to address challenges big and small across the supply chain, from commodity to customer. Founded in 1974, HAVI employs more than 10,000 people and serves customers in more than 100 countries. HAVI’s supply chain services are complemented by the customer engagement services offered by our affiliated company The Marketing Store. For more information, please visit HAVI.com.
This is a hybrid role based at 345 N Morgan St, Chicago, IL 60607. Candidates must reside in the Chicago metropolitan area. Relocation assistance is not offered at this time.
As a Data Architect, you will be responsible for our overall Data Platform design and implementation, including data platform strategy, data modeling and data governance practices.
- Data Platform – Defining and managing frameworks and standards for our Data Platform based on Databricks Lakehouse
- Data Modeling – Designing and updating data models
- Data Governance – Establishing and implementing Data Governance policies including compliance, privacy, quality and security
Responsibilities:
Data Platform:
You will work closely with our Cloud Infrastructure team, our Enterprise and Solution Architects, and our DevOps team to build and maintain our Data Platform. You will be responsible for the Administration of our Data Platform to support all HAVI initiatives.
Key Responsibilities for the Data Platform:
- Definition and implementation of Data Platform infrastructure jointly with our Cloud Infrastructure team
- Definition and maintenance of our Unity catalog implementation including catalogs, foreign catalogs, schemas and other aspects such as Delta Sharing
- Backup and Recovery requirements and activities as related to our Data Platform
Data Modeling:
You will work closely with our data engineers, data analysts, data scientists, and business stakeholders to understand their data requirements and translate them into logical and physical data models. You will also be responsible for ensuring the accuracy, completeness, and consistency of our data models, as well as maintaining their documentation and metadata. Your expertise in data modeling tools and techniques, as well as your ability to communicate complex data concepts to non-technical stakeholders, will be critical to the success of our data-driven projects.
Key Responsibilities for Data Modeling:
- Business Needs Analysis: Understanding the organization's data requirements and translating them into effective data models
- Data Model Creation: Designing conceptual, logical, and physical data models using tools like Entity-Relationship Diagrams (ERDs) and data modeling software
- Database/Lakehouse Design: Defining how data is stored, accessed, and related within a database, data warehouse or Databricks Lakehouse
- Data Integrity: Ensuring data accuracy, consistency, and reliability through data quality rules and governance
- Collaboration: Working with database administrators, data analysts, and software developers to implement and maintain data models
- SQL Proficiency: Writing complex queries to retrieve and maintain data. Optimizing queries for a Databricks Lakehouse
- Performance Optimization: Tuning data models for optimal performance and scalability
- Documentation: Creating and maintaining documentation for data models, data flows, and data dictionaries
- Staying Updated: Keeping up with new technologies and best practices in data modeling
Data Governance:
You will work closely with our IT and Business teams to define and provide governance for our Data.
Key Responsibilities for Data Governance:
- Definition of Data Governance standards and practices in conjunction with our IT and Business leadership teams
- Compliance monitoring related to Data Governance
- Definition and monitoring of Quality definitions and practices related to Data
- Definition and monitoring of Security related to the Data Platform in conjunction with our Security and Cloud Infrastructure teams
Qualifications:
- Experience: 6+ years experience in data modeling, Databricks Lakehouse design, data engineering, and data governance
- Strong Analytical Skills: Ability to analyze complex business requirements and translate them into data models
- Proficiency in Data Modeling Tools: Experience with ERwin, PowerDesigner, or other data modeling software
- Database and Data Platform Knowledge: Understanding of relational databases, SQL, data warehousing and Databricks Lakehouse concepts
- Data Integration: Experience designing and developing data integrations with various methods and tools including APIs
- Query and ETL/ELT Knowledge: Able to create and maintain complex SQL queries for performant data management. Knowledge of Python, PySpark and other scripting languages for development and automation
- Machine Learning and AI Integration: Experience with structuring databases and building pipelines that support AI and ML workloads
- Communication Skills: Ability to communicate effectively with both technical and non-technical stakeholders
- Problem-Solving Skills: Ability to identify and resolve data modeling issues
- Global Team: Able to work with teams from different time zones and review/supervise data engineering teams
- Data Security: Able to define and ensure compliance of data security standards within the company and when sharing data with external parties
- Bachelor's degree: In computer science, information technology, or a related field
*Starting Salary is $125,000-$135,000 with a 10% targeted bonus
Nearest Major Market: Chicago