Foundational Data Automation
Tata Consultancy Services / WNS
Analyst ProgrammerThe Narrative
"Eliminate human error and maximise efficiency in high-stakes, multi-client data analysis environments — turning what was a slow, manual, error-prone process into a reliable, automated pipeline that could scale across millions of records and dozens of simultaneous project streams."
Joined TCS/WNS as a junior analyst programmer and immediately took on a portfolio spanning multiple clients and multiple concurrent projects — market research datasets and medical questionnaire analysis with rigorous validation standards.
Developed a reusable suite of Shell Scripts that automated the full pipeline: raw data ingestion, schema-specific cleaning and transformation, multi-dimensional validation, and formatted output delivery. Scripts were parameterised by client schema so a single pipeline codebase served every account. Ran processing asynchronously overnight, freeing working hours for exception handling and quality verification rather than manual data entry.
Within months, productivity metrics ranked consistently at the top of the team — a pace sustained for five consecutive months. The automation patterns built here were adopted as the team standard. Before the first year was complete, transitioned from individual contributor to leading a team, applying the same systematic thinking that drove personal output to coordinate and grow a broader group.
Achieved the highest productivity hours in the organisation for five consecutive months while concurrently managing multiple client accounts and project streams with millions of records. Promoted to team lead within the first year.
Automation pipelines reduced error rates from 15% to 0.2% and cut report turnaround from days to minutes — establishing a new benchmark for data processing reliability that the wider team adopted as standard practice.
Strategic Decisions
Shell Scripting for High-Volume Data Pipelines
Repetitive, manual analysis of millions of records across market research and medical questionnaire datasets was prone to human error, took days per cycle, and did not scale across multiple simultaneous client accounts.
Developed a suite of reusable Shell Scripts to automate the cleaning, transformation, validation, and ingestion of raw data — parameterised per client schema so the same pipeline could service multiple accounts without duplication.
Turnaround time for major reports decreased from days to minutes, enabling simultaneous delivery across multiple client projects with 100% data consistency.
Multi-Client Parallel Project Management
Managing deliverables across multiple clients and project streams simultaneously created scheduling conflicts and the risk that quality would degrade under volume pressure.
Built a personal workflow system for batching and sequencing tasks by client SLA priority, with automated pipeline runs scheduled to run overnight so analyst time was reserved for validation and exception handling rather than raw processing.
Sustained the highest productivity hours in the organisation for five consecutive months across all active client accounts, with no missed deadlines and no client escalations.
Strategic Impact
Technical Ecosystem
Rooftop
Architecting for Exponential Growth
About·Contact·Privacy Policy·© 2026 Wenceslaus Dsilva