Data Architect - MCE

  • Bengaluru
  • Cgi
Position Description: Data architect , responsible for the design, implementation, and maintenance of the data infrastructure Metadata managementData life cycle managementData modellingData designing and Testing TechniquesPredictive Modeling/analyticsData Miningdata pipelines;on-premises and cloud storage facilities -- data lakes, data warehouses, data hubs;data streaming and Big Data analytics solutions (Hadoop, Spark, Kafka, etc.);machine learning and deep learning models; andbusiness intelligence tools.Data management and reporting technologies, Data VisualizationScripting : Python /Java/ PerlAI and ML tools: Vertex AI / Biqquery MLMonitoring and logging for ETL processSchema designskillset: Spark, Dataflow, SQL, Kafka, Pub/Sub, Hadoop and experience with batch/streaming jobs, -Multi cloud (GCP /Azure )-CI/CD-Azure data components (Azure Data Factory , Azure Databricks , Azure Fabric, Azure Purview, Power BI, Azure SQL)-GCP data components (Pub/sub, Dataflow, Dataproc, Airflow)-Hadoop (HDFS, HIVE, Impala, Scoop, Flume, Kafka, StreamSets, HBase-Scala, Spark, Java, Python -Kubernetes, Docker, ETL-Apache Trino/Impala/Sqoop/Hive-Cloud Security – Apache Ranger/ AD LDAP/ Could IAMRole:·Design, plan and deliver Solution Presentations, Demonstrations, Proofs of Concept and BenchmarksDocument Solution strategy and lead the technical Solution discussion with Stakeholders·Contribute technical responses for Requests for Information (RFIs) and Requests for Proposal (RFP)·Azure/GCP with specific deep skills in at least one or more of the following areas :- Application Infrastructure,-Application Development,-Data & AI and Security.If you are passionate about Azure/GCP Data and AI services, possess excellent communication and technical skills, and enjoy working in a dynamic and fast-paced environment, we would love to hear from you. Join our team and contribute to the success of our customers by driving the adoption of Azure/GCP Data and AI technologies. Skills: Apache KafkaApache SparkAzure Data FactoryData EngineeringGoogle Cloud PlatformHadoop HiveKubernetesPython