Summary
Overview
Work History
Education
Skills
Accomplishments
Additional Information
Certification
Timeline
SoftwareEngineer
Akram Khan

Akram Khan

Senior Data Architect Manager
Dubai

Summary

Big data analytics and platform architect with 14+ years of experience in data architecture & designing, data engineering, data analytics solutions, data security, data modeling, quality control, policy & governance, delivery & deployment of end-to-end cloud-native solutions. Assist the team on Business Analysis, requirements gathering, project Management, Project Estimation. Led, define and implement end-to-end modern data platforms, delivered edge to AI big data solutions on public/private cloud with Iaas/Paas/Saas services. Led team of 20+ data engineers, data analysts and data scientists to deliver end to end analytics projects, controlled budgets for improving operation process and generate ROI, consulting per-sales. Profound knowledge of big data technologies/Eco-systems, fluent with Cloudera, Apache Spark, Kafka, SQL, Apache Flink, Apache Airflow etc.

Overview

14
14
years of professional experience
6
6
Certifications

Work History

Senior Data Architect Manager

Unifonic
11.2021 - Current
  • Design data analytics platform architecture, define data governance and security policies, implement data mesh guideline to build data as a product
  • Deploy end to end data analytics solution, create the requirement analysis, platform selection, design of the technical architecture, application design and development, testing and deployment of the proposed solution
  • Consolidate teams deliverables to one efficient, cohesive unit; supporting the Chief Architect, Delivery of technical strategy & road-map
  • Profound knowledge of big data technologies/Eco-systems, fluent with Hadoop, Spark, Kafka, SQL, Apache Flink, MPP DB,
  • Data Lake, DWH, modern data platform
  • Clients: Banking/Finance, Telco and CPAAS
  • Design data analytics platform solution architecture, define data governance & security, data lineage, data catalogue, metadata management framework and practices
  • Coordinating with Client business/technical team to analyze existing system challenges/bottlenecks and propose most feasible solutions based on various big data technologies
  • Support migration plans of data from existing Data Warehouse to Data Lake-house or Modern data platform
  • Design the data model to be resilient to business demand changes and provide consistent performance, focus in all the critical areas and also venture into new areas, domains, data analytics and services support
  • Responsible for leading system integration and managing enhancements whilst providing functional expertise in digital transformation and functional & technical design for all services & solutions
  • Define the data storage platform, data science tools and frameworks and follow the in-country data regulatory compliance guidelines
  • Evaluated business requirements/use cases and prepared detailed specifications that follow project guidelines required to develop written scripts to handle data from various data sources
  • Assisted a cross-functional team to build Data application for near real-time operational analysis and reporting, moving operational barriers and empowering the business users with improved decision making
  • Led the transformation of existing data analytic infrastructure and data warehouse to a highly scalable, flexible, and cloud-native approach using appropriate big data open source technologies and frameworks
  • Enable and creation of a clear strategic technology road map to deliver overall business objectives and innovation of new business line using real-time Big Data /predictive analytics and data visualization
  • Contribute to new product validation & implementation, best practice recommendations, data Ingestion, data Access & data storage, application deployment, disaster recovery associated architecture, data replication and RTO/RPO.

Senior Big Data Architect

DXC
01.2021 - 10.2021
  • Design data analytics platform solution architecture, define data governance & security policies,implement data mesh guideline to build data as a product
  • Deploy end to end data analytics solution, create the requirement analysis, platform selection, design of the technical architecture, application design and development, testing and deployment of the proposed solution, per-sales experience
  • Clients: Banking/Finance, Energy domain
  • Design data analytics platform solution architecture, define data governance & security, data lineage, data catalogue, metadata management framework and practices
  • Excellent hands-on experience in developing Data Pipelines, Data Mapping & delivering large-scale Data-Science
  • Projects
  • Led team of data engineers, DataOps engineers, data analytics and data scientist for successfully project deliver
  • Successful phased implementation of Analytical Model Management Framework for complex Data Science life cycle on Cloudera data Platform for 2 big banks/finance sector in UAE
  • Deployed 2PB production Cloudera data platform cluster
  • Coordinating with Client business/technical team to analyze existing system challenges/bottlenecks and propose most feasible solutions based on various Big Data technologies
  • Support migration plans of data from existing Data Warehouse to Cloudera data platform cluster.
  • Implemented end to end data security where we have integrate AD, Ranger, Knox, TLS/SSL and Atlas to achieve it
  • Design the data model to be resilient to business demand changes and provide consistent performance, focus in all the critical areas and also venture into new areas, domains, BIG DATA and data services support
  • Deployed 18 nodes production Kafka cluster and setup complete data pipeline for real time and batch processing.

Senior Big Data Solution Architect

Edgematics
09.2019 - 12.2020
  • Design data analytics platform Architecture, define data governance and security frameworks, implement data mesh guideline to build data as a product
  • Deploy end to end data analytics solution, create the requirement analysis, platform selection, design of the technical architecture, application design and development, testing and deployment of the proposed solution
  • Clients: Telco, Banking & Finance and AI/ML domain clients
  • Design data analytics platform solution architecture, define data governance & security, data lineage, data catalogue, metadata management framework and practices
  • Led team of data engineers, DataOps engineers, data analytics and data scientist for successfully project deliver
  • IOT, AI and ML platform deployment and implementation with end to end data security and governance
  • Digital Transformation, Digital Culture Change, Machine Learning, Predictive Analytics, Performance Management,
  • Executive Coaching
  • Data Lake, Data Hub and modern data platform setup using Cloudera Enterprise data center solution and cloud native solutions
  • Experience in drafting solution architecture framework that rely on API’s and microservices
  • Experience in cloud native solutions and AI services (AWS, AZURE andGCP)
  • Set up real time and batch data processing platform using Cloudera (CDH) platform along with Confluent Kafka,
  • RHEL Open Shift and CICD pipeline for complete automate the upstream and downstream code released management
  • Expertise in setup big data security in data center solution and cloud big data native solution
  • Design and Implemented the data warehouse solution using Snowflake
  • Migrate the on-premise MIS and data warehouse to Snowflake on public cloud.

Senior Technical Lead DataOps

12.2018 - 09.2019
  • Edge to AI data analytics solution, creating the requirements analysis, the platform selection, design of the technical architecture, application design and development, testing and deployment of the proposed solution
  • Deployed 200 nodes Hortonworks cluster to handle and process 4 PB data, setup real time stream pipeline using Kafka cluster of 18 nodes and dynamic dashboard with 16 nodes Druid cluster
  • Client: Banking/Insurance/Food Industries/ Retails/Pharmaceutical companies
  • Led the transformation of existing data and analytic infrastructure to a highly scalable, flexible, and performing big data platform using appropriate big data open source technologies
  • Developed Spark SQL to process data from HIVE tables and saving results to MongoDB
  • Managing and deploying big data enterprise solutions, Hadoop/Spark, Kafka, ELK, Druid clusters, Integrating EDW with Data lake to setup modern data lake solutions
  • Been a part of Big Data as a service (BDAAS) build new Hadoop clusters and Hadoop deployment decisions
  • Successful phased implementation of Analytical Model Management Framework for complex Data Science life cycle on Hadoop Eco system for one of the biggest US Pharmaceuticals clients
  • Coordinating with Client business/technical team to analyze existing system challenges/bottlenecks and propose most feasible solutions based on various Big Data technologies
  • Hadoop Cluster Security setup with Kerberos, Sentry/RANGER, KNOX, KTS/KMS/Ranger KMS, TLS/SSL
  • Set up HDFS ACL’s with Sentry & Ranger for authorization, Knox for client gateway request, SSL/TLS implementation for securing HTTP traffic
  • AD/LDAP integration with Ambari/Cloudera for SSO and Ranger/Hue user and group sync
  • Hadoop architecture and configuring various components such as HDFS, YARN, MapReduce (MR1 & MR2), Tez,
  • Sqoop, Flume, Pig, Hive/Impala, Zookeeper, Oozie, Ranger, Knox, Sentry, Kafka, Storm, Solr, HBase, HDF, Apache
  • NiFi, Apache Zeppelin etc
  • Implement high availability for NameNode, Resource Manager, HiveServer2, Impala, Hue with HA-Proxy and
  • Zookeeper/JQM
  • Commissioning/Decommissioning of Node on running cluster
  • Capacity planning and re-sizing
  • Setup new Hadoop users, setting up Kerberos principals and validating their access
  • Importing and exporting data using Apache Sqoop from HDFS to Relational Database System and vice-versa
  • Analyze multiple sources of data ingestion and data structure whether it is structured, unstructured or semi structure data to propose design data architecture solutions for scale-ability, security and high availability
  • Set up data pipeline with Kafka, NiFi, Sqoop and Flume
  • Integrate Kafka with NiFi and Spark
  • Contribute to HDP/HDF cluster architecture design review, validation & performance optimization as required
  • Contribute to new product validation & implementation HDP Security guidance, best practice recommendations &
  • Implementation Data Ingestion, Data Access & Data storage Application Deployment, Disaster recovery associated architecture; data replication; RTO/RPO
  • Responsible for providing thought leadership emanating in the form of discussion papers and white papers to create best in class practices and tools for all client engagements

Senior Big Data Consultant

CenturyLink
05.2017 - 12.2018
  • Coordinating with Client business/technical team to analyze existing system challenges/bottlenecks and propose most feasible solutions based on various Big Data technologies
  • Consulting pre-sales, Consolidate all team deliverables to one efficient, cohesive unit, Ownership of all technical output; supporting the Chief Architect, Delivery of technical strategy, road-map
  • Client: Banking/Insurance/Food Industries/ Retails/Pharmaceutical companies
  • Responsible for implementation and support of the Hadoop Cluster environment (CDH, HDP and Pivotal)
  • Led the transformation of existing data and analytic infrastructure to a highly scalable, flexible, and performing big data platform using appropriate big data open-source technologies
  • Evaluated business requirements/use cases and prepared detailed specifications that follow project guidelines required to develop written scripts to handle data from various data sources
  • Assisted a cross-functional team to build BIG Data application for near real-time operational analysis and reporting, moving operational barriers and empowering the business users with improved decision making
  • Enable and creation of a clear strategic technology road map to deliver overall business objectives and innovation of new business line using real-time Big Data /predictive analytics and data visualization
  • Installation, configuration, supporting and managing Hadoop Ecosystems (Hive, Impala, Zookeeper, Oozie, Hue etc)
  • Implemented High end HDP security with KERBEROS with RANGER and KNOX integrated with LDAP
  • Analyze multiple sources of structured and unstructured data to propose and design data architecture solutions for scale-ability, high availability
  • Clearly able to identify data patterns and design/implement as per solution needs
  • Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments
  • Working with data delivery teams to setup new Hadoop users
  • This job includes setting up Linux users, setting up
  • Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users
  • Cluster maintenance as well as creation and removal of nodes using tools like Ganglia, Nagios, Cloudera Manager
  • Enterprise, Dell Open Manage and other tools
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.

Senior Big Data Consultant

Accenture
05.2017 - 12.2018
  • Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability
  • Subject matter expert in key Data analytics platform components that support the end to end data supply chain: data ingestion, storage, processing and exploitation
  • Responsible for defining and policing the Data analytics platform technology guardrails within which spokes must operate
  • Collaborating with Architect and supporting Project and providing specialist expertise within the DRA team
  • Implemented High end HDP security with KERBEROS with RANGER and KNOX integrated with LDAP
  • Responsible for driving a culture of relentless continuous improvement within DRA
  • Drive proactive to build the pipeline, using specialized knowledge and skills to prospect, qualify, deliver, and optimize the platform
  • Administration and Monitoring Hadoop clusters in both environments such as Development Cluster, Production
  • Cluster HDP and CDH with the help of Cloudera Manager,Ambari
  • Set up and manage Name-node High Availability with journal node and Zookeeper Implementation
  • Implemented automatic failover zookeeper and zookeeper fail-over controller
  • Commission/Decommission failed and new node in the cluster
  • Administer Kerberos Security on Hadoop, create key tab files, principal and set appropriate permissions for the principals per requirements
  • Analyze multiple sources of structured and unstructured data to propose and design data architecture solutions for scale-ability, high availability
  • Clearly able to identify data patterns and design/implement as per solution needs
  • Aligning with the systems engineering team to propose and deploy new hardware and software environments required for Hadoop and to expand existing environments.

Senior Big Data and Cloud Consultant

IBM, Australian Bank
10.2013 - 09.2015
  • Implementation and support (post implementation) of the enterprise Hadoop environment
  • Capacity planning and performance tuning
  • Work closely with infrastructure, network, database and application teams
  • Monitoring and managing storage that is used by Hadoop system
  • Configuring security for Hadoop Cluster
  • Monitor Hadoop cluster connectivity and performance
  • Manage and analyze Hadoop log files
  • Developed and document best practices
  • Support Hadoop Ecosystem components and third-party s/w using in Hadoop environment
  • Work with vendors (Cloudera) to resolve issues
  • Helping big data developers on big data infrastructure related issues.

Senior Big Data Hadoop/UNIX Administrator

HCL
12.2010 - 09.2013
  • Big data Hadoop Administration and Monitoring Hadoop clusters in both environments such as Development Cluster,
  • Production
  • UNIX Administration responsibilities
  • P2V migration and automation.

Senior Big Data Hadoop/UNIX Administrator

FirstSource
06.2009 - 11.2010
  • Installation, configuration & troubleshooting of Linux Servers
  • New Server Build and OS Hardening
  • Apache Hadoop and Ecosystem installation, configuration and troubleshooting
  • User Creation (Local OS user)
  • Hadoop Job monitoring and Cluster health check and generate reports
  • Monitoring the system performance during peak hours and suggest and recommend upgrading if bottlenecks occurs
  • User and group service Management (Creation, Deletion and modification)
  • Filesystem Management and Process monitoring
  • Installation, configuration & troubleshooting of Linux Servers
  • Storage management and monitoring the resources' usage periodically.

Education

Master of computer application (MCA) -

RGTU (Rajiv Gandhi technical university)

Bachelor of Computer Application (BCA) - undefined

Barkatullah university

Skills

Modern Data Architecture & Platform Solutions

undefined

Accomplishments

  • AWS Solution Architect
  • AZURE Solution Architect
  • Certified Kubernetes Admin and Developer
  • Cloudera Data Platform Administrator
  • Vertica Data Platform Specialist
  • Confluent Kafka Specialist
  • Container & Kubernetes Essentials with IBM Cloud
  • Hortonworks Data Platform Operations: Security -36367 -DEL 570
  • Data Governance and Security Professional

Additional Information

  • MPP DB, Data Lake, DWH, Cloudera, Address aspects such as data privacy & security, data ingestion & processing, data storage & compute, analytical & operational consumption, data modeling, data virtualization, self-service data preparation & analytics, AI enablement, and API integration.

Certification

Cloudera Certified Administratior

Timeline

Senior Data Architect Manager

Unifonic
11.2021 - Current

Senior Big Data Architect

DXC
01.2021 - 10.2021

Senior Big Data Solution Architect

Edgematics
09.2019 - 12.2020

Senior Technical Lead DataOps

12.2018 - 09.2019

Senior Big Data Consultant

CenturyLink
05.2017 - 12.2018

Senior Big Data Consultant

Accenture
05.2017 - 12.2018

Senior Big Data and Cloud Consultant

IBM, Australian Bank
10.2013 - 09.2015

Senior Big Data Hadoop/UNIX Administrator

HCL
12.2010 - 09.2013

Senior Big Data Hadoop/UNIX Administrator

FirstSource
06.2009 - 11.2010

Master of computer application (MCA) -

RGTU (Rajiv Gandhi technical university)

Bachelor of Computer Application (BCA) - undefined

Barkatullah university
Akram KhanSenior Data Architect Manager