Bigdata Architect
Description
Job description
As a Data Architect you will be responsible to defining the data architecture strategy, standards and patterns. You will bring expertise in developing data governance, integration and analytics strategies. As a data architect you will contribute to critical system architecture decisions and guide multiple teams.
Details on tech stack
RESPONSIBILITIES:
• Collaborate with cross-functional teams for their data modelling, engineering, science, and analytical requirements.
• Define Data Governance strategy for enterprise data model development, data security, and privacy project work.
• Ensure data flows and related information are being tracked, managed, and documented.
• Model integrative business views and determine the quality-of-service requirements — data currency, availability, response times and data volumes.
• Establish and enforce data standards, naming conventions and patterns.
• Define data cleansing, wrangling, enrichment, transformation, ingestion, extraction, streaming processes.
• Document data models and the data ecosystem at large using tools of various means.
• Defining data sharing strategy internally and externally providing secure data access guidelines.
• Design, develop and maintain database schema object including tables, indexes, partitions and clusters for database applications and data warehouse systems.
• Define processes to handle sensitive data in a multi-tenant environment.
• Adhere to all industry standards, procedures, and guidelines, including audits.
• Create data mapping documents and work on industry interoperability of data.
• Maintain database performance by identifying problems and control cloud cost consumption.
4. Min requirements to the candidate
· 10+ years of Data Architecture experience as well as process designing and experience.
· 5+ years of hands-on experience leading large-scale global data cloud warehousing, analytics projects and mission-critical data pipelines and features using modern big data architectures
· 5+ years of experience with GCP services such as Big Query, Cloud SQL, Cloud Composer, Data Flow, Spanner.
· Experience with Big Query is a must, Snowflake good to have.
· Retail and Supply Chain industry experience is a must.
· Proficiency with Data Modeling experience and tools (e.g., Power designer, Sparx, Erwin)
· Proficiency with complex SQL and NoSQL (MongoDB - Document, Cassandra - Wide Column, Graph) experience
· Proficiency & experience with designing customer centric database models, datasets and warehousing data modeling.
· Experience in advanced Big Data concepts including Streaming Analytics, Data as a Product.
· Experience with Reporting Technologies (e.g., Tableau, Looker, PowerBI)
· Knowledge of different schema and file formats e.g. XSD, EDI, Parquet, Avro, JSON
· Experience with Predictive analytics, AI/ML and tools such as Databricks, Vertex AI
· Experience with Business Requirements definition and management, structured analysis, process design, use case and technical documentation.
5. Nice to have requirements to the candidate
GCP must. Other cloud exposure good to have.
• Collaborate with cross-functional teams for their data modelling, engineering, science, and analytical requirements.
• Define Data Governance strategy for enterprise data model development, data security, and privacy project work.
• Ensure data flows and related information are being tracked, managed, and documented.
• Model integrative business views and determine the quality-of-service requirements — data currency, availability, response times and data volumes.
• Establish and enforce data standards, naming conventions and patterns.
• Define data cleansing, wrangling, enrichment, transformation, ingestion, extraction, streaming processes.
• Document data models and the data ecosystem at large using tools of various means.
• Defining data sharing strategy internally and externally providing secure data access guidelines.
• Design, develop and maintain database schema object including tables, indexes, partitions and clusters for database applications and data warehouse systems.
• Define processes to handle sensitive data in a multi-tenant environment.
• Adhere to all industry standards, procedures, and guidelines, including audits.
• Create data mapping documents and work on industry interoperability of data.
• Maintain database performance by identifying problems and control cloud cost consumption.
4. Min requirements to the candidate
· 10+ years of Data Architecture experience as well as process designing and experience.
· 5+ years of hands-on experience leading large-scale global data cloud warehousing, analytics projects and mission-critical data pipelines and features using modern big data architectures
· 5+ years of experience with GCP services such as Big Query, Cloud SQL, Cloud Composer, Data Flow, Spanner.
· Experience with Big Query is a must, Snowflake good to have.
· Retail and Supply Chain industry experience is a must.
· Proficiency with Data Modeling experience and tools (e.g., Power designer, Sparx, Erwin)
· Proficiency with complex SQL and NoSQL (MongoDB - Document, Cassandra - Wide Column, Graph) experience
· Proficiency & experience with designing customer centric database models, datasets and warehousing data modeling.
· Experience in advanced Big Data concepts including Streaming Analytics, Data as a Product.
· Experience with Reporting Technologies (e.g., Tableau, Looker, PowerBI)
· Knowledge of different schema and file formats e.g. XSD, EDI, Parquet, Avro, JSON
· Experience with Predictive analytics, AI/ML and tools such as Databricks, Vertex AI
· Experience with Business Requirements definition and management, structured analysis, process design, use case and technical documentation.
5. Nice to have requirements to the candidate
GCP must. Other cloud exposure good to have.