ABHIPSA BISOI
Big Data Engineer
Experienced Big Data Engineer with 5.5 years in designing and maintaining large-scale business applications focused
on data migration, integration, conversion, and testing. Skilled in SQL, Hadoop, and Big Data technologies, with
3.5years of hands-on experience using tools like Sqoop, HDFS, Hive, Apache Spark, and AWS. Proven track record
in delivering scalable, real-time data solutions, seeking to apply this expertise in a business application role to support
organizational growth.
Contact Professional Experience
GRID DYNAMICS(October 2024 to Present)
• abhipsa.bisoi01@gm Project Undertaken: Worked as Big Data Developer for Visa
ail. com Roles/Responsibilities:
• 8895052974
➢ Created Data Lake by consumer data from various sources and expose it to
• Bangalore, Karnataka
multiple systems.
Skills ➢ Extracting data from AWS data warehouse onto Spark Data Frames.
Data Eco System ➢ Created EC2 instances and EMR clusters for development and testing.
• Hadoop ➢ Performed step execution in EMR clusters for the spark job deployment
• Sqoop as per requirements.
• Apache Spark ➢ Loaded data onto Hive from Spark RDDs and data frames for further processing.
• Hive
• PySpark ➢ Worked with broadcast variables.
• AWS ➢ Queried data using SparkSQL on top of Spark engine for faster datasets
Languages processing.
• SQL ➢ Created multiple Hive tables, running hive queries in those data, implemented
• Scala Partitioning, Dynamic Partitioning and Bucketing in Hive for efficient data access.
• Python ➢ Processed web URL data using Scala and converted it to data frames for
• Unix Shell Script further transformations.
Version Control ➢ Generated complex JSON data after all the transformations for easy storage and
• GitHub access as per client requirements.
Distribution ➢ Storing the spark processed data in HDFS/S3 with appropriate file formats.
• Cloudera 5.12
➢ Regular performance tune of Hive queries to improve data processing and
Operating System retrieving.
• Linux
• Window
Databases INTELLISENSE SOFTWARE TECHNOLOGIES(Sept 2021 to Aug 2024)
• SQL Server Project Undertaken: Worked as Spark-SQL Developer for Blue Dart
• MYSQL
Roles/Responsibilities:
• NOSQL
➢ Performed Import and Export of data into HDFS and Hive using Sqoop and
managed data within the environment.
Strength ➢ Involved in creating Hive tables, data loading and writing hive queries and was
responsible for Optimizing Hive queries that helped in saving Cost to the project.
• Technical and
business ➢ Managed Hive Tables and created child tables based on partitions.
expertise.
➢ Involved in working on the Data Analysis, Data Quality and data profiling for
• Exp. in cost
handling the business that helped the Business team.
cutting for
better profits.
➢ Loaded and transformed large sets of semi structured data and understands the
• Able to make Complex Data Processing needs of big data.
fast POCs for a
concept. ➢ Applied knowledge of Hive schema evolution, SchemaRDD, Spark partitioning, Spark
integration, Spark memory tuning, and Spark performance tuning throughout the
• Taking proper
project to efficiently handle complex Spark workloads and improve data processing
requirement performance.
from clients.
➢ Development of Code&peer review of assigned task and Bug fixing and Have done
POCs using Apache Spark SQL.
EDUCATION ➢ Used Agile methodology to work with IT and business team to progress efficient system
development.
[Link](EEE)
N.I.S.T ACCEND SYSTEMS(JULY 2019 TO AUGUST 2021)
Project Undertaken: Worked as Hadoop Developer for Hershey’s
ODISHA
Roles/Responsibilities:
2011-2015
➢ Executed seamless data import and export operations between HDFS and Hive, utilizing
Sqoop, while adeptly managing data within the environment .
➢ Processed data from REST APIs, showcasing proficiency in managing intricate JSON
structures and transforming them into structured data .
➢ Engaged in managing schema evolution using AVRO, ensuring smooth adaptation to
evolving data structures .
➢ Proficiently loaded and transformed large volumes of semi-structured data,
encompassing formats such as XML, JSON, Avro and Parquet .
➢ Coordinated and obtained necessary approvals from diverse clients for all proposed
changes, ensuring a collaborative and efficient workflow.
➢ Showcased expertise in troubleshooting application and server errors, swiftly resolving
issues to maintain optimal system functionality.