John I. Lee *********@*****.***
********@*****.***
OBJECTIVE.
Details-oriented, precise, and resourceful computer software professional with extensive experience developing sophisticated and innovative cost-reduction and productivity solutions to complex business and IT problems for numerous Fortune 500 companies, seeking an IT consultant position as an application programmer.
TECHNICAL SKILLS
Operating Systems:
Windows 95/98/2000/NT/XP, HP-UNIX, AIX, Solaris, PC DOS, OS/390 IBM MVS, UNIX (Bourne/Korn shell scripting, VI, EX)
Programming Environments:
SAS (Base, Macro, Stat, Graphs, UNIX-PC-Mainframe), Ab Initio (ETL, GDE, EME, Co>Operating System), Teradata (MLOAD, FASTLOAD, BTEQ/ITEQ, TPUMP, TPT, FASTEXPORT, SQL Assistant, SAS/ACCESS interface to Teradata, DB2),
Teradata Table Compression (MVC, ALC, BLC),
Teradata(DBQL(logtbl,steptbl,objtbl,sqltbl,explaintbl),Viewpoint,SWS,TVI,BAR,SLES),SAS DDE(dynamic data exchange), Oracle (SQL*Plus, PL/SQL, SQLLOAD, ADO, SAS/ACCESS to Oracle), Cognos, SAS/AF (Frame, SCL, Components, Attributes w/dot notation, SCL to Frame, Properties, Form/Table Viewer, SAS base),SAS Grid Manager,
SAS Metadata Server (Platform Job Scheduler for SAS includes Platform LSF (an execution agent). Load Sharing Facility (bsub, bmod, bkill, bstop, bresume, bbot, btop, bjobs, bhist, bpeek, bqueues, and bparams)
SAS Graphs/Gchart (block, hbar, hbar3d, vbar, vbar3d, pie, pie3d, donut, star),
Metadata Driven Programming Technique using SAS Macros (parameters) and data steps on validation and imputation on user inputs. .
Internet, OOP, MS-Office, Miscellaneous Technologies:
MS-Office (Access, Excel, Word), ODBC HTML, Cold Fusion/IIS, JAVA (J2SE), JavaScript, VBScript/ASP, Visual Basic 6.0, MS SQL Server, TCP/IP, FTP, Lotus Notes, Outlook, PeopleSoft Tools, SQR, NEON-Shadow Direct, Microsoft SMS,Q&E, ESSBASE, Microfocus COBOL, SQL, Microsoft VB, VBA (Visual Basic for application) utilized macros, sub procedures, function procedures, properties, methods on Excel).
Mainframe Technologies:
DB2 (SQL, QMF, SPUFI, DCLGEN, COBOL, SAS/ACCESS to DB2), FOCUS, COBOL/COBOLII, Panvalet, Command Level CICS, JCL/PROCs, TSO, ISPF, SPF, PDF, IBM Utilities, Abend Aid, IBM Report Writer, Easytrieve/Easytrieve Plus, IDCAMS, VSAM/KSDS, ICETOOL/SYNCSORT, FileAid/Dumper
EDUCATION
B.S. in Computer Science, Mathematics
CONTINUING PROFESSIONAL EDUCATION
SAS / Teradata Certified.
SOLEX Academy E-Commerce Program – Web-related applications development using HTML, DHTML, VBScript, SQL, JavaScript, ASP, ADO, OOP, JAVA (J2SE, AWT, Swing)
LAN\MIND – UNIX
CARA – Professional Systems Analyst Development
Chicago Educational Center – Advanced SQL, SQL*PLUS, PL/SQL, Oracle 7 Applications Tuning
PeopleSoft Corporation – PeopleSoft nVision Reporting, PeopleSoft General Ledger I, PeopleTools I
Microsoft – Visual Basic
Arbor – ESSBASE
PROFESSIONAL EXPERIENCE
PureTech Information Services, Inc. Oct 2013 - Present
Senior Consultant Contractor.
Provide and support variety high effective techniques
(SAS, ETL, Teradata, UNIX, Windows, Mainframe) for various organizations.
Support many financial organizations such as Citi group,Discover, Capital one under PureTech Information Services, Inc,
My responsibilities was design, planning new and update, collaborate old Teradata system which contains many (100+) tables as well as creating many new tables under different database system.
Manage/Monitor system performance, including workload management and throttling in TASM, query performance and tuning, physical node behavior, and
Amp distribution Consult on physical database implementations, DML, DDL, PPI, skew analysis, join methods, and data access paths.
Strong knowledge of Teradata Architecture: Nodes (AMP &PE), VProcs, BYNETS
Efficient usage of Teradata DDL Statements – Derived Tables, Volatile Tables and Global Temporary Tables
Hands on Knowledge on Teradata Concepts – Primary Index, Partitioned Primary Index, Secondary Index, Join Index and Compression Techniques.
Understand the Concept of Locking Mechanisms in Teradata
Experience on Query Tuning mechanisms using indexes & proper usage of Statistics in Teradata
Utilized Teradata (MLOAD, FASTLOAD, BTEQ/ITEQ, TPUMP, TPTexport, TPTload, FASTEXPORT, SQL Assistant.
Under new database system tested, debugged, reported to management, implemented and mentor others.
Utilized SAS/Access interface to connect new Teradata database to retrieve and pull data into SAS platforms.
Enjoyed to exam an inefficient existing SAS program and find places that enhancements could be made, quantify and justify the enhancements, make the changes, test, debug, implement them.
Abbott/AbbVie – Waukegan, IL April 2013 – Oct 2013
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
Utilized DBQL to trace inefficient SQL for Suspect Query Report.
Teradata (v13)/SAS(9.3) application developer, responsible for AMP(access module processor) which is main core in the Teradata RDBMS and the Virtual Processor used to manage the database, handle file tasks and manipulate the disk subsystem in the multi-tasking and possibly parallel-processing
Created tables allow Unique and Primary keys as defined constraints on a table, so the rows across the AMPs evenly distributed and shared across the pre-defined number of AMPs.
Utilized Hash Function to verify even data distribution over all AMP’s.
Created Teradata RDBMS support seven types of Indexes (UPI, USI, NUPI, NUSI, join index, Hash index, PPI) for good query performance.
TERADATA utilities TPT,Tpump(5 DML),Fastload,Multiload(5 DML) Fastexport, BTEQ, analyze Performance analysis, SQL Optimization, Teradata SQL Assistant.
SAS 9.3, 9.2, SAS EG 4.3, 5.1, Teradata 13.10., Linux, UNIX, Windows7 (PC) Environment.
Support PPD (pharmaceutical products) create many SAS support programs for end-users to make decisions on Promotional Marketing Data both in SAS and Teradata.
Worked on Healthcare Claims and SAS/Base, SAS/SQL, SAS/Macros, SAS/Stat, SAS/EG, SAS/PROCs.
Utilized SAS/Access interface Pass-through (PROC SQL w/LIBNAME) to database (Teradata 13.10, DB2, Oracle).
Extract historical data from data warehouse(Teradata 13.10) into SAS 9.3,9.2 environments
And redesign data to transpose vertically or horizontally for users.
Utilized ‘proc import’ to manipulate Excel (csv, tab, dlm) to create SAS datasets to feed into SAS process.
Utilized ‘proc export’ for SAS datasets to create Excel (tab, dlm, csv) format.
Merge many SAS files to consolidate into one central location.
Created many high-level SAS Macros and utilized SAS Array’s to improve upon and simplify existing unorganized data processes. Used positional and key- word parameters to process Macros.
ARRAYs and DO loops - do over, do until, do while, and do index DIM, multiDIM.
SAS/Base, SAS/AF (Frame), SAS/SQL, SAS/Macros, SAS/Stat (summary, means, freq), SAS/EG (4.3, 5.1), SAS/PROCs proc report, proc summary nway chartype autoname auto label, proc means, proc tabulate with colpctn, colpctsum, pagepctn, pagepctsum, reppctn, reppctsum, rowpctn and rowpctsum, PCTN,PCTSUM.
Analyzed high volume of SAS datasets and report abnormal cases.
Create many SAS Macro jobs within Macro job and Automated in Unix used AT,Crontab command.
Utilized SAS Grid manager client utility to process large amount of data as quickly as possible for data-intensive projects.
Base SAS ODS (html, rtf, pdf, csv, csvall, proclabel).
Utilized SAS Enterprise Guide submit SAS programs to the Grid with options of megaserver, metaphor.
Utilized SAS Enterprise Guide generate ODS output on the Grid.
Utilized ‘proc ANOVA’ analyzed variance for balanced data using Model statement with dependent variable given independent variable.
Utilized ‘proc REG’ statistical tech. to estimated/analyzed relationships among variables it include modeling and help understand values changes on dependent variable on given independent variables.
Utilized ‘proc gchart’ to generate data chart reports.
Utilized Metadata Driven Programming Technique using SAS macros (parameters) and data steps.
Utilized ‘CDISC (Clinical Data Interchange Standards Consortium) procedure’ to import an XLM document that conforms to CDISC ODM version.
CAPITAL ONE – Mettawa, IL April 2012 – December 2012
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
Teradata (v13)/SAS(9.3) application developer, responsible for AMP(access module processor) which is main core in the Teradata RDBMS and the Virtual Processor used to manage the database, handle file tasks and manipulate the disk subsystem in the multi-tasking and possibly parallel-processing
Utilized DBQL to trace inefficient SQL for Suspect Query Report.
Created tables allow Unique and Primary keys as defined constraints on a table, so the rows across the AMPs evenly distributed and shared across the pre-defined number of AMPs
To improve query performance Created Teradata RDBMS support seven types of Indexes (UPI, USI, NUPI, NUSI, join index, Hash index, PPI).
Utilized Hash Function to verify even data distribution over all AMP’s.
Created many high-level SAS Macros and utilized SAS Array’s to improve upon and simplify existing unorganized data processes. Used positional and key- word parameters to process Macros.
ARRAYs and DO loops - do over, do until, do while, and do index DIM, multiDIM.
SAS/Base, SAS/AF (Frame), SAS/SQL, SAS/Macros, SAS/Stat (summary, means, freq), SAS/EG (4.3, 5.1), SAS/PROCs proc report, proc summary nway chartype autoname auto label, proc means, proc tabulate with colpctn, colpctsum, pagepctn, pagepctsum, reppctn, reppctsum, rowpctn and rowpctsum, PCTN, PCTSUM.
Analyzed high volume of SAS datasets and report abnormal cases.
Create many SAS Macro jobs within Macro job and Automated in Unix used AT,Crontab command.
SAS 9.3, 9.2, SAS EG 4.3, 5.1, Teradata 13.10., Linux, UNIX, Windows7 (PC) Environment.
Utilized Teradata utilities to convert Mainframe flat files, SAS dataset and Excel (.csv) into Teradata environment
Generated decision making customized SQL reports for business community
Utilized Teradata options such as TENACITY, SLEEP to avoid an abort when encounter the maximum number of Teradata utilities are running
Utilized Teradata SQL assistant ODBC-based client utility used to access and manipulate data
The Explain statement is used to aid in identifying potential performance issues On every queries.
TERADATA utilities TPT,Tpump(5 DML),Fastload,Multiload(5 DML) Fastexport, BTEQ, analyze Performance analysis, SQL Optimization, Teradata SQL Assistant.
CITI GROUP – Elk Grove Village, IL May 2011 – December 2011
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
Utilized DBQL to trace inefficient SQL for Suspect Query Report.
Teradata (v13)/SAS(9.3) application developer, responsible for AMP(access module processor) which is main core in the Teradata RDBMS and the Virtual Processor used to manage the database, handle file tasks and manipulate the disk subsystem in the multi-tasking and possibly parallel-processing
SAS/Base, SAS/AF (Frame), SAS/SQL, SAS/Macros, SAS/Stat (summary, means, freq), SAS/EG (4.3, 5.1), SAS/PROCs proc report, proc summary nway chartype autoname auto label, proc means, proc tabulate with colpctn, colpctsum, pagepctn, pagepctsum, reppctn, reppctsum, rowpctn and rowpctsum, PCTN, PCTSUM.
Created many high-level SAS Macros and utilized SAS Array’s to improve upon and simplify existing unorganized data processes. Used positional and key- word parameters to process Macros.
ARRAYs and DO loops - do over, do until, do while, and do index DIM, multiDIM.
Utilized SAS data view, contains a definition of data that is in a shared environment within organization
Wrote many korn shell scripts that executes SAS programs to create numerous SAS tables for business analyst
Handle Card holder files, Account master files for Sears, Oil companies, Home Depot and Retails
Used many SAS stored procedures in conjunction with ‘data step’ process
Procedures SQL, sort, freq, format, CNTLIN, CNTLOUT, summary/autoname, compare, compare two SAS datasets for discrepancy, transpose, print, report, contents, datasets, import, export, append/force.
Create functional specifications and perform validation of reports
Strong SAS Developer with business analysis skills that can analyze code and suggest solutions for automation and process improvements. Able to articulate with the business users as well as IT project team.
Created Teradata RDBMS support five types of Indexes(UPI, USI, NUPI, NUSI, join index
Based on business needed for higher level of performance and efficiency and used data management such as indexing, joining, sorting, aggregation
Used pass-through to connect Teradata from Unix bring Teradata data into SAS/Unix environment
Created/manipulated account credit master related tables in SAS/Unix environment for users
Loaded SAS tables into Teradata used utilities such as Fastload, Multiload, BulkLoad,
TPT (Teradata Parallel Transporter)
Generated customized SQL reports for business community
Utilized SAS/Access interface to TERADATA database via pass-through to extact millions rows data using join options. inner join, left join, right join, full join.
Used TERADATA utilities Multiload, Fastload to create Teradata Tables with proper Primary Indexes
To improve query performance, utilized USI,NUSI,PPI, HASH INDEX and JOIN INDEX.
The Explain statement is used to aid in identifying potential performance issues On every queries.
TERADATA utilities TPT,Tpump(5 DML),Fastload,Multiload(5 DML) Fastexport, BTEQ, analyze Performance analysis, SQL Optimization, Teradata SQL Assistant.
DISCOVER FINANCIAL SERVICES / DISCOVER CARD – Riverwood, IL August 2006 – December 2010.
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
SAS 9.2 (Base, Macros, ODS, Stat, Access), Linux, UNIX (AIX, Korn Shell, VI editor), Teradata (v12), Oracle, Ab Initio, Cognos, Mainframe. SAS GRID environment
Utilized SAS/Access interface Pass-through(PROC SQL w/LIBNAME) to database(Teradata, DB2, Oracle) in UNIX and MainFrame environment
Used SAS, SQL and Ab Initio to extract and transform data from a wide variety of enterprise data sources – a large-scale
Utilized DBQL to trace inefficient SQL for Suspect Query Report.
Teradata Data Warehouse (tables with billions of rows), an Oracle Data Mart, conventional and partitioned flat files – on the UNIX (AIX, Korn shell), PC and Mainframe platforms.
Utilized SAS Grid manager client utility to process large datasets processing. Command-line utility.
Base SAS ODS (html,rtf,pdf,csv,csvall,proclabel).
Strong working knowledge of Unix and Shell Scripting
Good Mainframe knowledge with SAS extensive data manipulation experience
Automated scoring code initially developed by in-house statisticians and outside vendors to allow it to be more easily run on an ongoing basis, using shell scripting and SAS macros to eliminate error-prone manual intervention and provide required audit reports
Automated Reage and Performance systems and converted Python programs to SAS programs used SAS macro functions and UNIX’s CRON and AT command to setup automatic job schedule and generated adhoc report for users
Automated Direct payment systems used SAS macro functions and UNIX’s CRON and AT command to schedule the jobs and generated adhoc report for users
Procedures SQL, sort, freq, format, CNTLIN, CNTLOUT, summary/autoname, compare, compare two SAS datasets for discrepancy, transpose, print, report, contents, datasets, import, export, append/force.
Create functional specifications and perform validation of reports
Strong SAS Developer with business analysis skills that can analyze code and suggest solutions for automation and process improvements. Able to articulate with the business users as well as IT project team.
Analyzed high volume of SAS datasets and report abnormal cases.
SAS/Base, SAS/AF (Frame), SAS/SQL, SAS/Macros, SAS/Stat (summary, means, freq), SAS/EG (4.3, 5.1), SAS/PROCs proc report, proc summary nway chartype autoname auto label, proc means, proc tabulate with colpctn, colpctsum, pagepctn, pagepctsum, reppctn, reppctsum, rowpctn and rowpctsum, PCTN, PCTSUM.
TERADATA utilities TPT,Tpump,Fastload, Multiload,Fastexport, BTEQ, analyze Performance analysis, SQL Optimization, Teradata SQL Assistant.
Provided IT and data sourcing support for business analysts and statisticians in the Risk Management Department
Development environment was SAS 9.1.3 (Base, Macros, ODS, Stat, Access),Linux, UNIX (AIX, Korn Shell, vi editor), Teradata, Oracle, Ab Initio, Cognos, Mainframe
Used SAS, SQL and Ab Initio to extract and transform data from a wide variety of enterprise data sources – a large-scale Teradata Data Warehouse (tables with billions of rows), an Oracle Data Mart, conventional and partitioned flat files – on the UNIX (AIX, Korn shell, vi editor), PC and Mainframe platforms
Supported statisticians' modeling efforts on both the front-end (sourcing and transforming data) and back-end (reporting results, QA). Responsible for tuning statistician's SQL and SAS code, reducing model scoring run-time in one case from days to under an hour
Responsible for the initial set-up and day-to-day operations and support of a Teradata-based reporting Data Mart for Discover Personal Loans, which was used enterprise-wide to track the growth of the product from launch to over $1 billion in outstanding loans. Also supported and maintained daily standard report processing. Applied changes to reflect product enhancements, including a conversion from an in-house loan-servicing product to an external vendor's banking product (FISERV)
Worked with analysts and statisticians to implement line management strategies, including the generation of production transactions used by IT to effect changes to cardholders' accounts (credit line increases and decreases), as well as the associated required reporting
Worked with Marketing and Risk to develop and run processing to supply Consumer and Business Card data to an outside vendor (Argus) on a monthly basis to enable Discover to participate in industry-wide competitive analysis studies
Worked with the Risk and IT groups to transition code from successful model and strategy tests to permanent production status, including a Tradeline model that earned the Discover employees a President's Award
Processed numerous quick-turnaround ad hoc reporting and data requests for Discover and its outside consulting firms, e.g., Credit Card reform legislation requests. Results were delivered in a variety of formats – flat files, SAS data sets, Teradata tables, Excel, etc
Supported analysts' efforts to explore new sources of data, including new scores, mortgage data, and enhanced Credit Bureau information. Worked with analysts to pull samples to provide to the Credit Bureaus and outside consulting firms, and created SAS data sets and Teradata tables with returned data for follow-up, in-house analysis
Provided mainframe storage usage reports that enabled Risk Management to identify unused archived data (disk and tape), allowing the department to dramatically reduce recurring monthly storage charges
As a Teradata(v13)/SAS(9.2) application developer, responsible for AMP(access module processor) which is main core in the Teradata RDBMS and the Virtual Processor used to manage the database, handle file tasks and manipulate the disk subsystem in the multi-tasking and possibly parallel-processing
Created tables allow Unique and Primary keys as defined constraints on a table, so the rows across the AMPs evenly distributed and shared across the pre-defined number of AMPs
Created Teradata RDBMS support five types of Indexes(UPI, USI, NUPI, NUSI, join index)
Based on business needed for higher level of performance and efficiency
and used data management such as indexing, joining, sorting, aggregation
Used pass-through to connect Teradata from Unix bring Teradata data into SAS/Unix environment and vreated/manipulated account credit master related tables in SAS/Unix environment for users
Loaded SAS tables into Teradata used utilities such as Fastload, Multiload, BulkLoad Generated decision making customized SQL reports for business community
Utilized SAS/Access interface to TERADATA database via pass-through to extactmillions rows data using join options. inner join, left join, right join, full join.
Used TERADATA utilities Multiload, Fastload to create Teradata Tables with proper Primary Indexes To avoid table skew for good performance
Utilized Teradata options such as TENACITY, SLEEP to avoid an abort when encounter the maximum number of Teradata utilities are running
Utilized Teradata SQL assistant ODBC-based client utility used to access and manipulate data.
To improve query performance, utilized USI,NUSI,PPI, HASH INDEX and JOIN INDEX.
The Explain statement is used to aid in identifying potential performance issues On every queries.
CITIGROUP RISK MANAGEMENT, Elk Grove Village, IL March 2006 – June 2006.
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
Utilized SAS/Access interface Pass-through(PROC SQL w/LIBNAME) to database(Teradata, DB2, Oracle) in UNIX and Mainframe environment
Defined, designed and developed systems to support Management. Also investigated, analyzed, documented and resolved system issues
Development environment included SAS 9.1.3, UNIX Solaris 5.9, vi editor, IBM Mainframe (MVS), PC and Microsoft products
Responsible for populating a large UNIX-based Reporting Data Mart (20+ GB, 1000 variables)
Supported ETL processes that loaded MSR (Mailed System Records) mainframe data into the SAS environment
Worked on the Historic Repository conversion, migrating processing from the TSYS credit card system to the FDR credit card system
Developed and tested SAS programs to maintain SAS datasets for the Account, Balance, Rewards and Transaction applications
Generated statistical analytics reports for Business Clients
Helped less-experienced programmers finish their programs and ready them for production implementation
Processed large-scale SAS datasets – e.g., 80 million observations, 2100 variables – using SAS indexing to improve performance and compression to reduce space requirements
Developed and tested numerous SAS programs and UNIX scripts, ensuring that the resulting processing was production-ready – return codes checked, variables verified for valid and meaningful values, UNIX scripts created for automated job scheduling and notification
Utilized custom SAS processing (DATA _NULL_, Merge), as well as a number of SAS PROCs, including Means, SQL, Summary, Forms, Print, Tabulate, Freq, Report, Rank, CORR, Univariate, Standard, Append, Explode, Export, Format, Registry, Release, Datasets, Compare, Import, Sort, Contents, Options, Source, Convert, Optload, Copy, Optsave, Transpose, Printto, Trantab
Employed SAS macros and arrays to help minimize coding redundancy
Tuned SAS programs to reduce CPU resource usage and boost performance.
Procedures SQL, sort, freq, format, CNTLIN, CNTLOUT, summary/autoname, compare, compare two SAS datasets for discrepancy, transpose, print, report, contents, datasets, import, export, append/force.
SAS/Base, SAS/AF (Frame), SAS/SQL, SAS/Macros, SAS/Stat (summary, means, freq), SAS/EG (4.3, 5.1), SAS/PROCs proc report, proc summary nway chartype autoname auto label, proc means, proc tabulate with colpctn, colpctsum, pagepctn, pagepctsum, reppctn, reppctsum, rowpctn and rowpctsum, PCTN, PCTSUM.
ABBOTT LABORATORIES (CHR BUSINESS SYSTEMS) – Abbott Park, IL February 2004 - December 2005.
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
Data management, data manipulation, statistical analysis and reporting using SAS 8, UNIX, DB2, Mainframe (MVS), PC and Microsoft products (Excel, Access, VBA, pivot tables, charting). Designed, developed, implemented, tuned and supported many SAS programs, SAS datasets and SAS macros
Worked on Healthcare Claims and SAS/Base, SAS/SQL, SAS/Macros, SAS/Stat, SAS/EG, SAS/PROCs.
Assisted internal business clients with their requirements gathering, design, development, construction, and implementation efforts
Supported the production Corporate Human Resource System – 300 SAS jobs in a mainframe, Web, UNIX and PC environment
Designed, developed, implemented and supported SAS-based New Hire and Transferred Employees Status weekly, monthly, quarterly and annual processing
Helped build company’s new HOSPIRA system (150 SAS jobs), which processed a variety of files and reports
Created 200 statistical and analysis SAS ad-hoc reports to satisfy various departments’ requirements for timely analysis
Developed Web-based Corporate Employee File Organization (CEFO) application in HTML and Cold Fusion that was used to maintain employees’ records on-line
Responsible for the support of 75 AIMS DB2 tables, which stored current and historical data for use by ad hoc requestors from various departments. Provided tech support for the system’s users
Mentored less-experienced employees, providing guidance on how to program and maintain jobs efficiently
Created Infopac (on-line report viewing system) and Dispatch tables
Maintained a number of COBOL programs used to populate DB2 tables with up-to-date information
Supported COBOL/UNIX data manipulation processes, including UNIX Bourne/Korn scripts, to handle feeds to/from various systems
Worked on a team that created large mainframe (MVS) flat files used to load the MENTOR system’s Oracle database tables
Extracted HR data using Tesseract in conjunction with COBOL/DB2
Utilized IBM DB2 catalog tables to identify prime keys for multiple-table joins so appropriate changes could be made to improve performance
Created many new mainframe PROCs with symbolic parameters
Loaded various DB2 tables using DB2 utilities; created many VSAM files (KSDS) with alternate indices and path entries
Used Excel and Access VBA to manipulate .txt, .csv and Excel files. Created pivot tables for analysis, charting, reporting purposes
Procedures SQL, sort, freq, format, CNTLIN, CNTLOUT, summary/autoname, compare, compare two SAS datasets for discrepancy, transpose, print, report, contents, datasets, import, export, append/force.
WALGREEN CO., Deerfield, IL January 2003 – January 2004
Senior SAS / Teradata developer
Support production system (Teradata, ETL, SQL,UNIX and SAS) process high volume (ten’s millions) data as input and output..
Performed requirements gathering, design, development, construction, and implementation for internal business clients
Supported legacy AP, DMS (Check Print System), FA (Future Rent) and Lawson AP Systems
Built Travel Meal System using MicroFocus COBOL (Express 3.1) in conjunction with Oracle Database via ODBC
Upgraded /modified various MicroFocus COBOL programs in the DMS system to improve efficiency and boost performance
Created and maintained UNIX Bourne shell scripts and COBOL programs used to process data feeds
Maintained and enhanced COBOL programs used to update DB2 tables
Created, maintained and tested mainframe MVS JCL and numerous SAS programs – SAS procedures used included PROC Summary, PROC Tabulate, PROC Means, PROC Access, PROC Report, PROC Freq, PROC SQL, PROC Sort, PROC Print, and PROC Forms
Used SAS Data _Null_ to convert SAS data to mainframe flat files and generate customized reports
Performed complex data manipulation against master and transaction files using SAS (update, merge)
Responsible for archiving SAS data to tape and disk backup for history purposes
CNA INSURANCE, Chicago, IL July 1990- February 2002
IT Application Specialist / Systems Analyst,
Supported the needs of a number of internal business clients throughout the entire systems lifecycle – requirements, design, development, testing, implementation, maintenance and support
Development was done in a variety of environments and platforms, including the mainframe (MVS, COBOL/ COBOL II), UNIX (HP), PC (Visual Basic), database (DB2, Oracle, Teradata), SAS (multiple platforms)
Assumed full responsibility for the Information Diversify Facility System (Oracle data warehouse), as well as the Claims Verification System (DB2 database, Visual Basic scripts)
Developed a number of on-line applications that utilized Visual Basic 6.0 on the front-end, COM/DCOM/MTS on the middle-tier, and UNIX/Oracle on the back-end. Created documentation for the systems, implemented them, and also trained users
Designed, wrote, tested and implemented a Claim/Policy Verification System that employed Visual Basic scripts which communicated with DB2 via ODBC. Provided documentation for the system and trained users
Implemented improved Visual Basic programs and forms to address users’ requests
Enhanced many COBOL/COBOLII programs, and made technical recommendations to project managers
Converted Visual Basic 3.1 16-bit applications to 32-bit Visual Basic 6.0 and implemented the upgraded applications. Provided VB mentoring and training to less-experienced team members
Built a large-scale Oracle data warehouse on HP-UNIX for CNA’s Monthly Financial Close
Generated status reports and documented problems and their resolution for team members and managers
Assumed a group DBA role