Dinesh Khanal
Dinesh Khanal
“100% Remote”
Expertise:- GCP Cloud data implementation, SQL, Python, Data Management, Big Data, Apache
Summary:
• Dinesh has 7+ years of experience in all the phases of the Software Development Life Cycle including design,
architecture, implementation, release, unit testing, refactoring and maintenance of web-based applications using
Big Data & Hadoop technologies.
• Expertise in major components in Hadoop Ecosystems like Spark, HDFS, HIVE, PIG, HBase, Zookeeper, Sqoop, Oozie,
Flume, Kafka.
• Experience with using Hadoop clusters, Hadoop HDFS, Hadoop tools and Spark, Kafka, Hive in social data and
media analytics using Hadoop ecosystem.
• Excellent understanding of Hadoop distributed File system and experienced in developing efficient MapReduce
jobs to process large datasets.
• Experience in developing SPARK applications using Spark tools like RDD transformations, Spark core, Spark
Streaming and Spark SQL.
• Experience in optimizing the data storage in Hive using partitioning and bucketing mechanisms on managed and
external tables.
• Proficient in migrating the data using Sqoop from HDFS to Relational Database System and vice-versa according to
client's requirement.
• Installation of various Hadoop Ecosystems and Hadoop Daemons.
• Experience data processing like collecting, aggregating, moving from various sources using Apache Flume and
Kafka.
• Ability to analyze search performance metrics such as query latency, throughput, and relevance scores to identify
and resolve performance bottlenecks.
• Strong collaboration skills to work effectively with cross-functional teams such as product management, software
engineering, and data science.
• Excellent communication skills to articulate technical concepts to both technical and non-technical stakeholders.
• Hands-on experience extracting data from different databases and risk management.
• Hands-on experience with installing Kerberos Security and setting up permissions, set up Standards and Processes
for Hadoop based application design and implementation.
• Experienced in cloud architecture using AWS ecosystem and GCP and Distributions (Cloudera Hadoop and
Hortonworks Hadoop).
• Worked on Azure Databricks, Azure Data Factory and Azure Data Lake.
• Experienced in Risk Frontier for risk management and data analysis.
• Experience in Big Data platforms like Hadoop platforms Microsoft Azure Data Lake, Azure Data Factory, Azure
Databricks, Azure Blob Storage and Graph Databases.
• Brief exposure in Implementing and Maintaining Hadoop Security and Hive Security.
• Experience integrating Kafka and Spark by using Avro for serializing and deserializing data, and for Kafka producer
and consumer.
• Expert in writing complex SQL queries with databases like MySQL, Oracle, SQL Server.
• Understanding of Hadoop Big Data architectures, data movement technologies, database partitioning, database
optimization, and building communication channels between structured and unstructured databases.
Education:
• Masters in engineering management, San Jose State University, San Jose, CA
• GCP Certified Google Cloud Engineer.
Technical Skills:
Hadoop Ecosystem: HDFS, MapReduce, YARN, Hive, Pig, Impala, Zookeeper, Sqoop, OOZIE, Kafka, Sqoop, Flume, Spark,
Pyspark, AWS, GCP, Azure
Languages: C, Java, SQL, PL/SQL, Scala, Shell Scripts
Operating Systems: Linux, UNIX, Windows
Databases: HBase, Oracle, SQL Server, MySQL, MongoDB, Cassandra
Cloud Services: AWS, GCP, Azure, Elasticsearch, Cloudera, Databricks, Hortonworks, Elastic, Cloud Foundry, Elastic Cloud,
EMR
Web Tools: HTML, JavaScript, XML, JDBC, EJB, MVC, AJAX, JSP, Servlets
Operating System: Windows, UNIX, Linux
Professional Experience: