Está en la página 1de 8

Competency Name: BigData and Hadoop Ecosystems

Broad Definition: This competency refers to the ability of the associate to work on Bigdata and Hadoop Ecosystem. It covers Mapreduce, HDFS,
Pig latin, Hive, Hbase & YARN, Apache Avro, DataFu, Apache Flume, Fuse-DFS, Apache HBase, Apache Hive, Hue, Apache
Mahout, Apache Oozie, Apache Sqoop, Apache Whirr, and Apache ZooKeeper. This competency measures skill levels of an
associate for applying Hadoop components to the Hadoop based solutions.
Criteria E0 E1 E2 E3
I Criteria for defining
proficiency vis-à-vis
knowledge & skills
1.Precondition Should have knowledge on: Should satisfy all the Should satisfy all the conditions of Should satisfy all the
[ Expressed as a prerequisite - • Programming - Core Java conditions of the previous the previous proficiency level conditions of the previous
in terms of academic • Conceptual understanding of proficiency level proficiency level
qualification AND / OR Linux and basic commands Should have Linux
industry experience (Linux OS E0) knowledge
(whichever is applicable) AND
/ OR relevant knowledge in
any other competency

TCS Confidential
2. Knowledge Should have : Should have: Should possess: Should possess:
[Expressed as Area of • High level understanding of • knowledge on various • Advanced Knowledge on the • Design & Architecture
knowledge and not skill, Concepts like Distributed File ecosystem components like integration of various components
aspects of atleast one
includes knowledge of Systems, Distributed Map/Reduce, Hive, Pig and of various Hadoop ecosystem distribution in High Availability
process, methodologies, Processing, Streaming & Hue using one Distribution components like hive, pig, spark,
cluster environment
concepts and tools.] Schema on Read e.g. Cloudera, Apache spark streaming, kafka, flume and
• Performance Tuning of
• Conceptual knowledge of • Ability to conduct PoC by its use cases. Hadoop ecosystem using
map reduce framework and its leveraging hadoop atleast one distribution
use in hadoop eco-system ecossystems like hive, pig, • Performance Tuning and best •Cluster sizing with knowledge
solutions like Pig and Hive. spark, spark streaming, practices for map/reduce jobs, of hardware architectures with
• Basic understanding of kafka, flume etc. hive queries, Pig and spark respect to failover and high
solutions which can be offered • Knowledge on writing unit programming paradigm in atleast availability.
on Hadoop Platform using its and integration test cases one distribution • Ability to compare scenarios
core attributes i.e. Adhoc- for hadoop applications for Hadoop implementation and
Querying using hive, using map-reduce and suggest the right fit
Algrothims and patterns using spark programming components for given customer
Map Reduce, High avaialbility paradigm. Ability to validate situation
options for data and / verify SQL procedures on • Ability to compare Hadoop
computing etc. Hive. distribution for customer
• knowledge on SQL situation and knowledge about
concepts and querying and more than 1 distribution
comparision to what • Ability to consult customer on
abilities are available on Hadoop implementation
hive.. strategy & the right product
fitment

TCS Confidential
Contd. • Knowledge of NoSQL, and
ability to create and defend
solution based on the kind of
use cases in business
porblem. Indepth knowledge of
One of various type of NoSQLs
avaialble i.e. Cassandra,
MongoDB etc. Should be able
to highlight solution around
consistency and availability.

3. Experience
[Expressed as duration in
months / years wrt exposure,
including breadth & the depth
of assignments; also beyond
project experience]

TCS Confidential
3.1 Type of project Should have: Should have: Should have: Should have:
experience • 0-6 months hands-on •Minimum 6 months hands- •Working experience in atleast •Design and development
[Includes 'type' in the form of experience on : on experience in atleast one distribution (Upto 1 yr of experience in atleast one
implementation, end-to-end one distribution experience, atleast two projects distribution (2 Year+ of
project, support, roll outs, •• Mapreduce, HDFS, Pig AND AND 2-3 PoCs) experience, atleast 3 projects
upgrades, professional latin, Hive, Hbase & YARN •Should have experience in •Should have hands experience in AND 2-3 PoCs)
services, consulting using atleast one distribution pseudo distributed mode performance tuning of atleast 1 AND
engagements AND / OR the setup distribution • Working experience in Java/
Number of projects AND / OR •• Apache Avro, DataFu, AND AND BI or similar projects in dev,
Time spent on the project ( for Apache Flume, Fuse-DFS, • Working experience in • Working experience in RDBMS design & architecture (5
eg., 3 consulting projects of Apache HBase, Apache Hive, RDBMS concepts or Linux concepts or Linux scripting (6 years+)
minimum 6 months duration ; Hue, Apache Mahout, scripting (Upto 6 months) months +)
1 roll out project of at least 18 Apache Oozie, Apache • Working experience in • Working experience in Java/BI
months] Sqoop, Apache Whirr, and Java /BI or similar projects or similar projects in dev, design
Apache ZooKeeper through in dev, design & & architecture (3 years+)
classroom hands on architecture (1 years+)
excercises and PoCs OR
•Working experience in
Hadoop (6 months to 1
Year of experience, atleast
one project OR 2-3 PoCs)

TCS Confidential
3.2 Beyond project N/A N/A Should have provided estimates Should have provided
development experience for Siebel UCM Application estimates for Siebel UCM
[Includes Configuration in MDM solution for Solution for 2 proposal
[1) Solutioning of proposals atleast 1 proposal OR
2) Reviews done OR Should have contributed 10
3) No. of contributions made Should have contributed 5 person- person-days of effort in a year
to trainings/training material] days of effort in a year in ISU Pre- in ISU Pre-sales involving
sales involving Siebel UCM Siebel UCM (Solution/
(Solution/ estimates) estimates)
OR OR
Should have done 1 Siebel Should have done 2 Siebel
CRM/UCM Project Technical UCM Project Technical Audit
Audit OR 1 Siebel CRM OR 1 Siebel UCM Review
Application Configuration OR
Technical Review Should have conducted 1
OR training on Siebel UCM
Should have conducted 1 training Application
on Siebel CRM Application Configuration(Siebel Essential
Configuration(Siebel Essential Training) for min 2 working
Training) for min 5 working days days
OR OR
Should have reviewed at least 2 Should have reviewed at least
implementation methodologies / 3 implementation
Process Artifacts / Knowmax methodologies / Process
Assets / Mighty Reusable (for Artifacts / Knowmax Assets /
Siebel UCM) Mighty Reusable (for Siebel
UCM)

4. Asset Creation
[Expressed as any reusable
and deployable assets
created in the knowledge
management system of the
organization]
4.1 Process frameworks N/A N/A N/A • Built a solution framework to
defined improve the productivitiy in
[Articulated as the process Hadoop projects
improvement designed OR a
new process framework
designed]

TCS Confidential
4.2 Tool creation N/A N/A Should have: Should have:
[Expressed as a system or • Contribution to least 1 • Mentored creation of at least
function tool designed OR deployable tool in any focused 1 deployable tool in any
created towards improving ecosystem component (e.g. code focused ecosystem component
business delivery / learning] review tool) (e.g. code review tool)

4.3 Reusable & deployable N/A N/A Should have: Should have:
assets created • Contributed to creating at least 1 • Mentored creation of at least
[Expressed as reusable and asset to help big data projects 2 assets to help big Data
deployable assets created (e.g. scripts, programs) projects (e.g. Best practices,
(like those in Mighty & scripts, programs, documents)
Knowmax]

OR
No. of training material N/A N/A Should have assisted in Should have created/updated
created creating/reviewing at least 1 at least 1 training material /
[Expressed as the number of training material / KSS KSS presentation pertaining to
relevant training material presentation pertaining to Siebel Siebel UCM Processes
created towards development UCM Application Configuration
of a particular body of Processes(per year)
knowledge or skill]
5. Industry contribution
[Expressed as the
contributions made to the
industry.This criteria should
be measured by mechanisms
for eg: TCS GEMS]

TCS Confidential
5.1 Recognized by Peers N/A • Should have responded to • Should have been recognized as • Should have been recognized
[Expressed as the recognition queries in Knowmax an SME for technical expertise as an SME for technical
given by the peer community discussion forum and (e.g. Customer appreciation for expertise (e.g. Customer
and should be measured by should be an active design & dev, Supervisor appreciation for design &
appropriate mechanism such member in communities in confirmation that associate can architecture, Paper
as TCS GEMS] Knowmax lead the team to design the presentation in TACTICS,
solution) Case study presentation in
• Should be an active member in ABIM, Solution Quality during
the discussion forums in the consulting engagements or
Knowmax and responding to proposals)
queries
Submitted patches to • Should be active member in
problems/issues in the opern Hadoop communities
source forums
• Should be an active member
in the discussion forums in
Knowmax and responding to
queries
AND/OR
Recognized by a Parent N/A N/A N/A Should have got certification
Body from hadoop distributer like
[Expressed as the cloudera, IBM etc
contribution recognized by a OR
renowned and recognized Should have published white
parent body] papers in Big data space in
recognized forums
OR
Submitted patches to problems
in the opern source forums

AND/OR
Invited as Speakers N/A N/A N/A Presented case study or
[Expressed as recognition solution in external Hadoop /
from an external body when it Big Data conferences like
extends an invitation to the Hadoop Summit
speaker to address a
recognized forum]

II Assessment / Evaluation
Methodology vis-à-vis
Proficiency Level

TCS Confidential
1. Certification As defined in the LO. As defined in the LO. As defined in the LO. As defined in the LO.
[Expressed as achievement of
external programs - certifying
a part of the proficiency level,
Achievement of internal
programs - internal trainings
attended]

2. Evaluation Test Should have completed the Should have completed the Should have completed the Should have completed the
[Expressed as a certification following: following following following
the individual has to achieve
towards the said proficiency Self assessment and Self assessment Self assessment Self assessment
level. (for eg.,Type of tests for Learning object. Learning object Learning object and Learning object
each proficiency should be Supervisor Assessment Supervisor Assessment
defined-like for E0 level it will SME Assessment
be only self assessment, for
E1 it should be panel review
etc)]

Version History
Project Details:
Unit: Analytics, Big Data and Information Management Services (ABIMS), DESS
Parent Project-
Project Owner/GL- Name & employee No : Kamlesh Mhashilkar, Pracitce Head, ABIMS

Amendment History:
Version Defined By Approved By - Role & Facilitated By Date
Group
Version 1 Vijay Garg (358943) Kamlesh Mhashilkar, Sulabha Manjrekar (384715) 29th April,
Manoj Mathai (175646) Pracitce Head, ABIMS Learning and Development 2015
G Ramsubbareddy (776037)
Rashmi Paranjape (325653)

TCS Confidential

También podría gustarte