The project is responsible for creating and providing a central knowledge graph for all the content. The knowledge graph is a technical representation of all the information from content assets and their relationships. They deal with various content types like movies, series, podcasts, music and more. Hence, a diverse set of data sources with varying semantics needs to be integrated. The resulting knowledge graph is made accessible for other products in the ecosystem via microservices and enables them to provide the best experience for the customers.
Scope of Service:
Consulting the data scientists in the development of algorithms and machine learning models es-pecially regarding productionization of these models Implementation of data pipelines for stream and batch processing that generate triples to feed our knowledge graph and that train and build our machine learning models
Design and implementation of microservices that provide REST and SPARQL APIs to query the knowledge graph, including tracing and monitoring
Deployment of the microservices in the productive cloud environment, considering scalability and high availability requirements
Implementation of CI/CD pipelines to deploy data pipelines and microservices to production
Requirements/Technology:
Google Cloud Platform, Terraform, GitLab
Kubernetes, Docker, GitLab CI/CD, ArgoCD, Dataflow, Apache Beam
BigQuery, BigTable
Python, SQL, Java
REST API, GraphQL