0% found this document useful (0 votes)
48 views3 pages

Manoj Kumar Arram Resume PDF

Manoj Kumar Arram is a Senior Data Scientist with over 9 years of experience, specializing in Generative AI, NLP, and machine learning applications. He has successfully led multiple projects, including developing tools for query recommendations, risk assessment, and document parsing, utilizing technologies such as Python, Spark, and AWS. His work has significantly improved operational efficiency and decision-making across various domains.

Uploaded by

nikoyami87
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
48 views3 pages

Manoj Kumar Arram Resume PDF

Manoj Kumar Arram is a Senior Data Scientist with over 9 years of experience, specializing in Generative AI, NLP, and machine learning applications. He has successfully led multiple projects, including developing tools for query recommendations, risk assessment, and document parsing, utilizing technologies such as Python, Spark, and AWS. His work has significantly improved operational efficiency and decision-making across various domains.

Uploaded by

nikoyami87
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

Manoj Kumar Arram

Senior Data Scientist


Bengaluru | 9441940141 | amanojkumar24@gmail.com
 stackoverflow.com/users/1559252/manoj-kumar  linkedin.com/in/arrammanojkumar/

Professional Summary Skills


Experienced Data Scientist with over 9 years of hands-on expertise in the field. Programming : Python, Java, R, Scala
Proficient in cutting-edge Generative AI concepts, specializing in the implementation
Data Science : Pandas, Tensorflow,
of RAG architectures and fine-tuning Narrow Language Models. Demonstrated
Keras, Scikit-learn,
success in developing scalable machine learning applications, leveraging tools such
Numpy, Matplotlib,
as Pyspark, Scikit-learn, and Pytorch to drive innovation in data-driven solutions.
Seaborn, Tableau
Extensive background in NLP, predictive analytics, and recommendation engines,
with a proven track record of delivering impactful projects. Skilled in collaborating Cloud. : AWS
with global teams to design and implement robust data pipelines using Pyspark,
Data Eng. : Spark ( Core, SQL,
effectively managing large volumes of structured and unstructured data.
Streaming), Kafka,
Adept at building scalable machine learning models using Spark, harnessing the
Hive, Cassandra, Pig,
power of Python libraries like Scikit-Learn, pandas, and numpy for efficient data
Oozie, YARN
wrangling and model development. Proficient in developing advanced modules for
tasks such as text summarization and Named Entity Recognition (NER), embedding No-Sql DB : Mongo, Cassandra,
intelligence into systems for enhanced functionality. Hbase
Recognized for developing predictive and classification models with high accuracy
Search Engine : Elastic Search
rates, meeting and exceeding business/domain expectations. Experienced in a wide
range of applications including predictive modeling, recommendation engines, sales Having Good experience with several
analytics, and forecasting models. Proficient in data wrangling, data visualization, IDE's like Jet brains, VS Code, Eclipse,
and driving actionable insights from complex datasets. And several versioning/tracking tools
like Git, JIRA

Project Details
Project #1: Ask Compliance
Professional Exp
Working in Standard Chartered
Application Description: Developed a tool to recommend similar queries from an
bank as a Manager-Data Scientist
existing corpus to assist advisory agents. This was achieved by generating text
from April 2021
embeddings and using a Vector Database for similarity search algorithms.
Worked as a Senior data scientist in
Ecosystem: Python, NLP, Vector DB (FAISS) Innominds Software Private Limited
from June-2015 to March 2021
Roles and Responsibilities:
Led the project from requirement gathering to production deployment.
Developed and implemented text embedding models to represent queries in a high-
dimensional space. Education
Integrated FAISS (Facebook AI Similarity Search) for efficient similarity search across Master Of Science - 2015
the corpus.
Optimized the system for quick and accurate recommendations to assist advisory IIIT-Hyderabad
agents. B Tech - 2012
Achievements:
JNTU-Hyderabad
Significantly reduced the time spent on each case by 3 hours, improving overall
Pre-University - 2008
efficiency and response times.
Enhanced the advisory process by providing accurate and relevant query AndhraPradesh State board
recommendations.

Project #2: Account Activity Review: Awards


Application Description: Developed a system to assign risk ratings to customers by Innostar of the month May 2016
analyzing their account transactions and transactional patterns. Innostar of the month May 2017
Innostar of the month Aug 2019
Ecosystem: Python, Machine Learning, Spark, Hive, Dataiku
Nominated to InnoStar of the Year
Roles and Responsibilities: in 2018
Led the entire application development process from inception to deployment.

Page 1 of 3
Designed and implemented machine learning models to assess customer risk based
on transaction history and patterns.
Integrated Spark and Hive for efficient handling and processing of large-scale data.
Utilized Dataiku for data preparation, model building, and deployment.
Achievements:
Significantly improved the efficiency of anomaly detection processes, reducing the
review time per account by 6 months.
Enhanced the accuracy of risk assessments, leading to better decision-making and
risk management.

Project #3: AI Pods


Application Description: Developed a bank-wide solution for parsing documents of
various formats. The process involved extracting text using OCR and parsing the output
for further analysis.
Ecosystem: Python, OCR, NLP.
Roles and Responsibilities:
Led the implementation of the document parsing process across the entire bank.
Utilized Natural Language Processing (NLP) techniques to parse and structure the
extracted text.
Integrated the solution with Spark for handling and processing large datasets.
Achievements:
Streamlined document processing workflows, resulting in significant time savings
and increased operational efficiency.
Improved data accuracy and accessibility by automating the extraction and parsing
of information from documents.

Project #4: Lenexa


Application Description: An IoT project focused on assessing patient health using data
from medical devices. The system analyzes patient risk by utilizing image processing and
deep learning algorithms.
Ecosystem: Python, Django, MySQL, Keras, TensorFlow, OpenCV (CV2), PIL
Roles and Responsibilities:
Led the entire application development process from inception to deployment.
Designed and developed deep learning and machine learning solutions for posture
classification, risk probability assessment, and risk timer.
Architected the data flow pipeline to ensure seamless integration and processing of
IoT data.
Developed and implemented the AI/ML modules of the application.
Achievements:
Successfully built a comprehensive system to monitor and assess patient health,
providing accurate risk evaluations.
Enhanced the reliability and efficiency of patient health monitoring through
advanced image processing and deep learning techniques.

Project #5: Match Light - Terbium Labs


Application Description: Developed a system to detect if any digital assets were
compromised on the Dark Web.
EcoSystem: Python, AWS, Machine Learning, Postgresql
Roles and Responsibilities:
Re-architected the entire backend process, achieving a 70% reduction in cloud costs
without compromising the detection of potential digital asset matches.
Led and managed the backend team, ensuring efficient and effective development
and maintenance of the system.
Optimized machine learning models and database queries to enhance performance
and reduce latency.

Page 2 of 3
Utilized AWS services to ensure scalable and cost-effective deployment and
maintenance.
Achievements:
Successfully reduced operational costs by 70%, significantly improving cost-
efficiency.
Maintained a high level of accuracy in detecting compromised digital assets,
ensuring robust protection for clients.

Project #6: Collective[I]


Application Description: A sales analytics product aimed at optimizing the performance
of sales and marketing teams. It provides various tools, utilities, and recommendations
using machine learning to enhance team efficiency and effectiveness.
Ecosystem: Spark, Python, Kafka, Hive, Impala, ElasticSearch, Neo4J, Scala, Scikit-learn,
Pandas.
Roles and Responsibilities:
Designed and developed Spark ML-based solutions to predict win/loss probability
and expected revenue of opportunities.
Architected and developed applications using PySpark to:
Track out-of-office (OOO) users.
Extract email signatures.
Mask secure information.
Enhance customer profiles by processing large volumes of structured and
unstructured data.
Extensively utilized the Stanford NLP library and wrote custom parsers for advanced
text processing tasks.
Designed a unification system using PySpark and MySQL to generate globally unique
identifiers (GUIDs), serving as the single source of truth for person and organization
identity across the application ecosystem. This module is the core of the entire CI
system.
Achievements:
Successfully developed and deployed machine learning models that significantly
improved sales and marketing performance predictions.
Enhanced data processing efficiency and accuracy, contributing to better decision-
making and customer insights.
Established a reliable unification system, ensuring consistent and accurate identity
management across all applications.

Project #7: Cogtax (Deloitte)


Application Description: Cognitive Tax is an indirect tax recovery product that utilizes
machine learning and OCR technology to help customers track overpaid taxes to the
government. The platform enables the tracking of thousands of invoices within minutes,
providing predictions on whether invoices are eligible for refunds.
Ecosystem: Regression, Clustering, Java, Spark, Microservices, Mongo DB
Roles & Responsibilities:
Developed machine learning classification models to classify the categories of
different taxation forms.
Primarily focused on modeling and fine-tuning machine learning algorithms for
accurate classification of taxation forms.
Collaborated with cross-functional teams to integrate machine learning models into
the Cogtax platform.
Contributed to the design and development of microservices architecture for
scalable and efficient processing of tax data.
Achievements:
Successfully implemented machine learning models that accurately classify taxation
forms, improving the efficiency of tax recovery processes.
Enhanced the Cogtax platform's capabilities to handle large volumes of invoices,
resulting in faster processing times and improved customer satisfaction.

Page 3 of 3

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy