Data Product Architect

Berkshire Hathaway GUARD

  • Parsippany, NJ
  • Permanent
  • Full-time
  • 30 days ago
ResponsibilitiesThe Data Product Architect plays a critical role in our Data and Analytics (DnA) Operating Model,contributing to the development of a Data Marketplace that provides valuable, reusable dataproducts for various business domains within the federated guild.In collaboration with Data Product Managers, this role defines and implements solutionarchitectures for data products that align with business objectives, value hypotheses, andanalytical use cases. The Data Product Architect is the custodian of technical assets, including:
  • Detailed design diagrams
  • Data contracts
  • Domain resources (e.g., infrastructure, compute capacity, storage)
Additionally, this role administers the CI/CD DevOps process, ensuring seamless deployment andcompliance with both domain-specific and platform-wide guardrails.Primary Objectives:
  • Build data products that support descriptive, diagnostic, predictive, and prescriptive analytics.
  • Utilize the DnA Platform’s services while allowing custom development within prescribed guardrails.
  • Drive business transformation by helping domains predict and optimize outcomes at scale.
Key Responsibilities:
  • Serve as a technical consultant for business owners and SMEs during the pre-data product development phase.
  • Lead the design and implementation of scalable data product architectures.
  • Ensure consistency between functional, service quality, and system management requirements.
  • Represent the data product solution design within the Federated Governance Guild.
  • Follow the DnA target architecture, leveraging and reusing existing patterns to align with core principles.
  • Create, iterate, and maintain technical documentation and data product architecture.
  • Optimize performance, cost-efficiency, and future scalability of data systems.
  • Act as a Microsoft Fabric expert, ensuring scalability, security, and governance of data solutions.
Technical Responsibilities:
  • Data Product Canvas Phase
  • Conduct data value chain analysis and assumption management.
  • Define solution requirements and establish high-level designs.
  • Data Product Delivery Phase
  • Develop detailed architectural designs.
  • Manage infrastructure, compute capacity, and storage.
  • Oversee CI/CD pipelines for data ingestion, transformation, and visualization.
  • Serve as an equal stakeholder alongside the Data Product Manager during user acceptance testing.
Leadership & Collaboration:
  • Mentor domain data engineers in best practices.
  • Drive data product development and delivery to achieve business goals.
  • Continuously assess data product usage and recommend enhancements.
  • Collaborate with the DnA Platform team to integrate new innovations.
Salary Range$100,000.00 - $170,000.00 USD per yearThe successful candidate is expected to work in one of our offices 3 days per week and also be available for travel as required.The annual base salary range posted represents a broad range of salaries around the U.S. and is subject to many factors including but not limited to credentials, education, experience, geographic location, job responsibilities, performance, skills and/or training.QualificationsMinimum Qualifications:
  • Bachelor’s or master’s degree in computer science, Information Science, or related field.
  • 6 to 8 years of hands-on experience in Data Product Architecture, with expertise in Data Mesh, Marketplace, and Product methodologies.
Key Technologies & Tools:
  • Microsoft Azure
  • Microsoft Fabric
  • OneLake/Delta Lake – Optimized storage for structured and unstructured data.
  • Apache Spark – Scalable framework for large-scale data processing.
  • SQL & T-SQL – Essential for data querying and transformations.
  • PySpark & Scala – Critical for big data processing and machine learning workflows.
  • Fabric Notebooks – Python, SQL, and Spark development environment.
  • Data Factory – ETL/ELT tool for data pipeline orchestration.
  • GraphQL API – Efficient querying and data integration across domains.
  • CI/CD Pipelines – Automating data workflows and deployments.
  • Livy API – Used for managing Apache Spark jobs.
  • Monitoring & Logging – Tools to track pipeline performance and health.
  • Fabric Advisor – Optimization tool for performance tuning.
  • Power BI – Advanced analytics and visualization.
  • VS Code Integration – Enables development and debugging of Fabric notebooks.
Soft Skills & Leadership:
  • Expertise in Agile software development methodologies.
  • Strong understanding of Data Product Management principles.
  • Ability to identify dependencies and solve complex challenges creatively.
  • Excellent communication and stakeholder management skills.
Preferred Qualifications & Industry Expertise:
  • Experience in P&C Insurance industry.
  • Microsoft Certified: Fabric Data Engineer Associate certification

Berkshire Hathaway GUARD