Post Job Free
Sign in

Senior Release/DevOps Engineer with 25+ Years Experience

Location:
San Jose, CA
Salary:
$145k/annually + Benefits
Posted:
November 24, 2025

Contact this candidate

Resume:

Saied Kamali

San Jose, CA, *****, USA

408-***-****

********@*****.***

https://www.linkedin.com/in/saied-kamali-35250684

Website Link:

https://672d22bc0804d.site123.me/

Abstract

Senior DevOps & Cloud Architect with over 25+ years of industry expertise

driving automation, infrastructure transformation, and release engineering. Adept

at architecting and managing scalable, secure, cloud-native environments using

AWS, Azure, Kubernetes, and Terraform, with deep experience in CI/CD

pipelines, IaC, and DevOps automation frameworks. Proven leader with a track

record of modernizing infrastructure, enabling digital transformation, and fostering

engineering culture rooted in innovation and excellence. Passionate about reducing

manual overhead via low-to-no-touch deployments and engineering-first

automation strategy.

Skills Obtained & Tools worked

With

Experience Gained

DevOps / Release Manager /

Configuration & Change Management

/ CI/CD Management, Release

management for web systems and

Content Management Systems (CMS)

25+ Years

CM Processes & Standards

18 Years

Unified Change Management (UCM)

15 Years

ClearCase

4 Years

ssPVCS/ Dimensions

2 years

Perforce P4D/NTX64

10 years

Team Foundation Server 2008-2013

6 years

StarTeam

1& half year

Visual Source Safe (VSS)

4 years

CVS/RCS

9 years

Java/J2EE

5+ Years

ANT

5 Years

ClearCase, Rational Unified Process

(RUP)

4 years

R&D Developer

6 years

C#

5 years

C/C++

6 years

AWS

5 years

Perl 5.0

10 years

Python 3.13.0a2

5 years

InstallAnywhere6.0

1 years

Visual Studio 2017/2019 Beta 2 & .Net

.NET Framework 3.5/4.x Beta 2,

Octopus 4.36.1 Deployment tool,

VSTS 2015

15 years

KSNT 8.1 / Cygwin Utilities

15 years

Windows PowerShell 5.1

6 years

CruiseControl.Net, JIRA, Hudson,

GIT 2.13.2. windows.1Windows.1

WebSphere, WebLogic, JBoss, Jenkins

2.5

5 years

Bash, shell, Cshell UNIX/Windows

Platforms

10+ years

Unix/Linux

10+ years

Octopus Tentacle - 5.0.12

4+ years

Octopus Deploy Server - 2019.12.5 LTS

4+ years

Octopus CLI - 7.1.3

4+ years

Salesforce, Copado DevOps Platform,

DevOps Azure, Octopus Deploy a

Deployment automation tool for

DevOps Teams

3+ years

Terraform v1.39 and Kubernetes

2 years

AWS Cloud Services; S3, RDS, EC2,

Lambada

5 years

Azure Cloud

5 years

Education:

Batchelor, Electronic & Electrical Engineering BEng (Hons) – Graduated 1983

Masters, Computer Science and Engineering/SCU – Graduated 1988

Certifications

Certification, Rational Corporation, Clear Case Meta-Data.

Certification, Rational Corporation, Clear Case Administration.

Certification, Rational Corporation, Clear Case fundamentals on UNIX.

Certification, Perforce Software Inc.p4 99.2 Administrator.

Certification, Starbase Corporation (Borland Software Corporation), StarTeam 4.0 Administrator.

Certification, Microsoft Web Development with Microsoft Visual Studio 2010

Web Design Theory and Best Practices Module 1 - Module 2

Microsoft System Center Designing Orchestrator – 2012

BizTalk Development/Deployment/Configurations

Windows PowerShell 2.0/3.0

Privacy and information Security Training Certificate

TGS – Practices & Delivery Information Security Training

Microsoft Azure Fundamentals Certification

Building Language Models on AWS

Advanced Testing Practices Using AWS DevOps Tools

AWS Command Line Interface (CLI) Basics

Getting Started with AWS on DevOps

AWS Compute Services Overview

Advance Testing Practices Using AWS DevOps Tools

Introduction to containers

Build and Deploy APIs with a Serverless CI/CD

AWS Skill Builder Course Completion Certificate

AWS Course3 Completion Certificate

AWS Course2 Completion Certificate

AWS Course Completion Certificate

Amazon API Gateway for Serverless Applications

Recent Experience

Onsite (San Jose, CA) July 2024-Present

I would like to share an update regarding my recent career developments. The Lockheed Martin projects I was involved with concluded on July 30, 2024, due to a lack of federal funding, and as a result, my MES and PLM team's contracts were not renewed. Taking this opportunity, I took some time off to travel overseas to my home country to attend to personal matters. I have recently returned to the U.S. and am actively re-engaging with the job market. Since then, I have received two offers but had to decline both due to differences in salary expectations. Additionally, I have participated in several first and second-round interviews with other organizations. I would welcome the chance to meet with you in person or have a conversation to share my credentials and to better understand your specific expectations and goals. Please let me know a convenient time for you. Thank you for your time and consideration. I look forward to the opportunity to connect.

Remote (Fort Worth, TX) Jan 2022 – July 2024

Senior Cloud DevOps Engineer/Hands on Software Configuration Manager

Client: Lockheed Martin (Lockheed Martin: Leading Aerospace and Defense)

Employed By: TEKsystems Inc.

Team: Agile Excellence

Accomplished/performed, Took ownership and maintenance of the following

activities at Lockheed Martin:

Responsible for the R&D, design, & implementation of new or modified software products like MES (Manufacturing Execution System), PLM (Product Lifecycle Management), PLM4ADP. Collaborate across Agile teams to gather their requirements, design, develop, implement, build, test, and maintain technical solutions. Lockheed Martin utilizes advanced Manufacturing Execution Systems (MES) and Product Lifecycle Management (PLM) systems to optimize production and manage the lifecycle of their complex aerospace and defense products. MES infrastructure

iBASEt MES Solution: Lockheed Martin's Aeronautics division, responsible for military aircraft design and manufacturing, has selected iBASEt's digital manufacturing suite as its next-generation MES. This solution aims to provide enhanced visibility, control, and efficiency in manufacturing operations.

Digital Continuity and Transformation: The MES deployment is a core part of Lockheed Martin's broader digital transformation strategy, aiming for seamless digital continuity across manufacturing engineering, process planning, shop floor execution, and quality management. The goal is to replace legacy systems and establish a single, integrated platform for a more adaptable and valuable approach to production.

Intelligent Factory Framework: Lockheed Martin is establishing an Intelligent Factory Framework, a cybersecure, standards-based network forming the basis of their Industrial Internet of Things (IIoT) solution. This framework connects manufacturing machines to enable real-time data collection and analysis, predictive maintenance, and optimized operations.

Worked closely with QA to ensure the end product is delivery and deployment to Prod is done with highest quality. Provide support and enhance Engineering productivity as needed. Used nexus3 GitLab AUTH plugin plugged GitLab. Scanners like Sona type Nexus Repository Pro were used to improve uptime with fast artifact availability, automatic failover and component replications.

Siemens Digital Industries Xcelerator PLM Portfolio: Lockheed Martin Aeronautics made a significant investment in Siemens' Xcelerator PLM portfolio to support their full lifecycle processes, from engineering to manufacturing to field support.

Teamcenter: Lockheed Martin has been using Teamcenter PLM software (part of the Siemens suite) for data management. This follows the adoption of Teamcenter by the U.S. Air Force as the standard for collaborative data management, influencing large defense contractors like Lockheed Martin.

Model-Based Enterprise (MBE): Lockheed Martin is transitioning to a Model-Based Enterprise (MBE) approach, leveraging the Xcelerator platform and other technologies to integrate digital models and the digital thread throughout the product lifecycle. This MBE strategy aims to improve efficiency, collaboration, and responsiveness to customer needs.

Nexus repository was being used to archive COTS (Commercial Off-The-Shelf software)

from third party vendor like Siemens.

Used terraform infrastructure which was defined both cloud and on-prem resources in human-readable configuration file, was an infrastructure orchestrator, which was addressing the needs to automate the lifecycle of environments.

Used TeamCity, for our build process which typically composed of multiple "build steps" which were individual tasks executed sequentially within a build configuration, each defined with a specific build runner to perform actions like compiling code, running tests, packaging artifacts, or deploying to an environment; essentially outlining the entire build lifecycle within a project.

Performed the key points about TeamCity build process steps:

Configuration within a build configuration:

Each set of build steps is defined within a specific build configuration within a TeamCity project, allowing for different build processes for different environments or deployment scenarios.

Build runners

Build runner is a part of TeamCity that allows integration with a specific build tool (Ant, MSBuild, Command Line, and so on). In our build configurations, a build runner was defined on how to run a build and report the results.

Sequential execution:

Build steps are executed one after another, with the next step only starting if the previous step completes successfully.

Customized:

we add, remove, or reorder build steps as needed to fit our project's build process. Typical build process steps in TeamCity:

Designed and implemented automated solutions for provisioning, configuring, and managing AWS services such Amazon Elastic Container Service (ECS), Amazon Elastic Kubernetes Service (Amazon EKS), REST API in API Gateway (REST).

Troubleshooted and resolved issues related to AWS services.

Worked on Migrating the infrastructure from the traditional AWS cloud environment to AWS GovCloud, for better security.

Managed applications in commercial Cloud providers, including Amazon Web Services AWS, GovCloud.

Implemented automated cloud systems using Lambda, resulting in 15% efficiency improvement.

Designed CloudFormation templates to build AWS services that supported enterprise applications.

Customized and configured AWS cloud environments, resulting in 20% more efficient workflows.

Assisted in designing and implementation of full data pipeline to support data queries and analysis in Cloud Data Warehouse in AWS GovCloud.

Monitor and optimize AWS services for cost efficiency.

Architect and manage infrastructure using tools like Kubernetes and Terraform.

Activities at LMCO with Kubernetes:

Architect and manage infrastructure using tools like Kubernetes and Terraform.

Kubernetes automates the deployment of containerized applications, allowing developers to focus on coding and testing rather than manual infrastructure management.

Kubernetes provides a standardized infrastructure for managing containerized applications, making it easier for developers, DevOps engineers, and other stakeholders to collaborate.

Kubernetes integrates well with CI/CD pipelines, allowing for automated testing, deployment, and rollback of applications.

Kubernetes provided an LMCO standardized infrastructure for managing containerized applications, making it easier for developers, DevOps engineers, and other stakeholders to collaborate.

Collaborate with product development, hosting, and QA teams to ensure alignment of DevOps strategies with business objectives.

Act as a hands-on mentor and guide for junior team members, leading by example to foster a culture of continuous learning and improvement.

Maintained CI/CD pipelines to automation of software delivery, data storage management, testing, and deployment processes.

GitLab CI/CD defined in configuration:

Workflow configuration files are written in YAML (Yet Another Markup Language) and are stored in the code's repository root.

Performed and supported GitLab-Jira integration.

Authored a playbook to support SCM/RM for a generic/LMCO specified for standard, procedures for development/other teams’ members.

Generated/maintained GitLab YAML (Yet Another Markup Language) config files/runners upon request, for deployment lifecycle for multiple applications pipelines, such as MES (Manufacturing Execution System), 3DX (At LMCO, "3DX" referred to the Dassault Systems' 3DEXPERIENCE platform), PLM (Product Lifecycle Management), PLM4ADP (Internal PLM) software, Aero IT Projects such as Data and Analytics, Inner resource and Engineering.

Implementing an MES (Manufacturing Execution System) typically involved a phased approach, starting with assessment and planning, followed by system selection, configuration, integration, and finally, rollout, testing, and ongoing optimization. This process ensures a smooth transition and maximizes the system's effectiveness. Here's a more detailed breakdown of the key performance steps:

Assessment and Planning:

Define Objectives

Process Analysis

Requirements Definition

Project Team Formation

Scope Definition

Infrastructure Planning

System Selection and Configuration:

MES Vendor Selection

System Design and Customization

Data Migration

Integration and Testing:

Third-Party System Integration

Hardware Configuration

Testing and Optimization

Rollout and post-implementation:

Change Management

User Training

Pilot Program

Full Deployment

Continuous Monitoring and Optimization

Post-Implementation Activities

The 3DEXPERIENCE platform facilitated Lockheed Martin’s digital engineering goals and help optimize the product engineering with an integrated platform approach. This approach lends flexibility to extremely complex programs and drives the advances in sophisticated aircraft innovation that are defining the 21st century of the aerospace industry.

PLM responsibilities:

Product strategy.

User experience design.

Agile PLM software.

Business analysis.

Marking Skills.

Leadership.

Communication Skills.

Collaboration.

Engineering expertise

The 3DEXPERIENCE (3DX) platform by Dassault Systems is widely used in the aerospace industry for various applications including Lockheed Martin where I performer my activities. Here are some key uses:

Design & Development: Lockheed Martin uses the 3DX platform to streamline the design and development of aircraft. It helps in creating detailed 3D models, simulating performance, and managing the entire lifecycle of the product.

Collaboration: The platform enables seamless collaboration among different teams and stakeholders, regardless of their location. This is crucial for large-scale aerospace projects that involve multiple partners and suppliers.

Digital Twin & Simulation: 3DX allows for the creation of digital twins—virtual replicas of physical assets. This helps in simulating and analyzing the performance of aircraft under various conditions, leading to better design and maintenance strategies.

Certification & Compliance: The platform aids in managing the certification process by ensuring that all regulatory requirements are met. This is particularly important in the aerospace industry, where safety and compliance are paramount.

Training and Simulation: Lockheed Martin uses the platform for immersive training solutions, blending real and simulated environments to enhance training effectiveness.

Used IaC (Infra-as-code) software tool using “Imperative approach” to provides a consistent command line interface (CLI) workflow to manage AWS Clouds.

The US federal government, including Department of Defense (DoD) and civilian agencies, often required our users to securely authenticate their identity and establish access controls to government networks, desktops, and other online resources using a Common Access Card (CAC)/Personal Identity Verification (PIV) card.

Used The National Institute of Standards and Technology (NIST) core functions. As developers for the mentioned apps navigate through the core functions, they gain a holistic understanding of how the NIST CSF can be applied to enhance security practices in the cloud. The functionalities are:

Identity

Protect

Detect

Respond

Recover

The network functions deployment is separated in two phases:

The required AWS infrastructure is created and managed via a central repository.

The configurations and codes were centrally stored in a GitLab repositories.

Used Azure DevOps Server 2022 exe version, in parallel for experimental purposes and possible consideration of replacement for current cloud being used.

Setup Azure Pipelines to automatically deploy to Azure Functions. Azure Pipelines allowed us to build, test, and deploy with continuous integration (CI) and continuous delivery (CD) using DevOps Azure.

There are now two versions of the Azure Function App task

AzureFunctionApp@1

AzureFunctionApp@2.

AzureFunctionApp@1 included enhanced validation support that makes pipelines less likely to fail because of errors.

Therefore, we used AzureFunctionApp@2 for our processes for the same purpose.

Used compliable languages and tools like C#, ASP.NET, JavaScript

Used IMS (incident management solutions) to streamline incident response communions by using IMS template. That helped us kick off the processes and alert our team with unexpected incident.

As soon as an incident is detected, an incident response team member filled out a simple form in Zapier Interfaces. They record details such as their email, incident name, description, status (investigating, identified, monitoring, resolved), and security level.

A workflow triggered, creating a dedicated Slack channel for the incident and tagging the relevant teams or groups.

The incident data is stored in a Zapier table and displayed below the form.

For IMS cybersecurity workflows Tines was used. Tines a powerful automation platform that excels our building cybersecurity workflows.

Use Cases:

Enriching threat intelligence.

Alerting for phishing attacks or suspicious login attempts.

Streamlining vulnerability management processes.

Automating endpoint detection and response workflows.

Onsite (San Mateo, CA) Oct 2017 – Jan 2022

Sr. App Developer & Sr. Build & Release Manager

Fisher Investments LLC.

Team: App Development Team

Accomplished/performed & maintained the following activities @ FI:

Applied Release management, for web systems and Content Management Systems (CMS), which was the systematic process of managing and delivering new or updated software components and applications. The objective was to efficiently and reliably deliver these changes to end-users while ensuring the stability and integrity of existing production environments.

Setup complete Build Environment including physical build server.

Meanwhile continued supports to development team with CI/CD on Non-official Build VMs.

Tools used to achieve the activities: GIT version 2.25.0. windows.1, TFS 12.0.40629.0,

PowerShell 3.0, 5.0, 5.1, GNU bash, version 4.4.12(1)-release (x86_64-pc-msys)

Octopus 4.36.1.

To describe our Octopus, Deploy is a sophisticated, best-of-breed continuous

delivery (CD) platform for modern software teams. Octopus offers powerful release orchestration, deployment automation, and runbook automation, while handling the scale, complexity and governance expectations of even the largest organizations with the most complex deployment challenges.

Our Octopus environment had a modern, friendly user experience that made it easy for our teams to self-service application releases, and democratizes how we deliver software. Octopus also has a comprehensive API - anything the UI can do; the API can do too.

Then Octopus takes over where our CI server ends, modelling the entire release orchestration process of software. This includes:

Release versioning

Environment promotion (beyond simple dev/test/prod workflows)

Deployment automation

Progressive software delivery (rolling deployments, blue/green, canary)

Configuration management

Approvals & ITSM integration

Deployment freezes

Coordinating deployments across projects and their dependencies

Complete customized scripting of CI/CD using PowerShell/Bash/Python 2.7.14.

Module written in C# / HTML /PowerShell to Support the Build automation processes.

Complete development of a UI to support internal-user self-build/run/status checks for commits/build/CI/CD phases.

Administered/Managed/maintained/supported TFS/Git/GitHub source control systems in paroral.

Performed migration & administration and Planning from TFS to GIT/GitHub by using git-tfs.

Administration & usage of GitHub for CI/CD GitHub Actions both allow you to create workflows that automatically build, test, publish, release, and deploy code. Maintenance of GitHub CI/CD and GitHub Actions share some similarities in workflow configuration:

Workflow configuration files are written in YAML and are stored in the code's repository.

Performed administration and migration from TFS (Team Foundation Server) to GIT & ultimately from TFS to GitHub.

Then migrated repositories from GitHub Enterprise Server to GitHub Enterprise Cloud. Using GitHub API.

About repository migrations with GitHub Enterprise Importer prerequisites:

Create our personal access token

Set up a migration source in GitHub Enterprise Cloud

Generate migration archives on our GitHub Enterprise Server instance

Uploaded migration archives

Started repository migration

Checked the status of the migration

Validated migration and check the error log

Administration of repository migrations performed with GitHub Enterprise Importer (GEI)

Prerequisites using CLI

Install the GEI extension of the GitHub CL

Update the GEI extension of the GitHub CL

Set environment variable

Set up blob storage

Generate a migration script

Migrate repositories

Validate our migration and check the error log

Used AWS clouds for code deployments to Prod.

Tried and initiated effective disaster recovery strategies with AWS, reducing potential data loss by 25%.

Deployed and managed AWS infrastructure using EC2, RDS and SES services, boosting uptime by 20%.

Managed company AWS usage, reducing costs by creating scalable solutions.

Used GitHub Actions which provided a straightforward way to set up CI/CD pipelines directly from our repository. Here are the key steps I took to create our own CI/CD pipeline using GitHub Actions:

Create a Workflow File:

In our repository, create & maintained a. GitHub/workflows directory.

Add a YAML file (e.g., ci-cd.yml) to define our workflow.

Specify the actions to be executed in response to specific events (e.g., pushes, pull requests).

Define Workflow Actions:

Within our workflow file, define the steps for our pipeline.

These steps can include installing dependencies, running tests, building artifacts, and deploying to our target environment.

Trigger Workflow on Events:

GitHub Actions can respond to various events, such as code pushes or pull requests.

Configure our workflow to trigger based on the events relevant to our project.

Test and Deploy:

As part of our workflow, include testing steps to ensure code quality.

Set up deployment actions to automatically deploy to our desired environment (e.g., staging or production).

Switched to Salesforce Copado API lifecycle stages; Planning, Building (CI/CD), Testing, Deploying, Monitory. Kept orgs in sync and prevent disruptive sandbox refreshes by automatically back-deploying changes to lower-level environments.

Selecting platform for Salesforce Metadata Pipelines, select a Git Repository, select branches working with, creating a Pipeline from the Pipeline tab was now set.

Creating a Pipeline from the getting staring page. Clone a Pipeline, ultimately creating Pipeline using the Pipeline wizard.

Downloaded Azure DevOps Server 2022.0.1 executable and Azure DevOps Server allowed the team to self-host the same modern dev services available in Azure DevOps, including boards (planning and tracking), Pipelines (CI/CD), Maintenance of repos (GitHub repositories and centralized version control), Test Plans (manual and exploratory testing), Artifacts (NuGet package management and pipeline artifacts), and more.

Used Azure Identity and Access Management (IAM) to enable framework that allowed me to manage access to Azure resources. It allowed me to control who has access to our resources and what they can do with them.

IAM was used to assign roles to users, groups, service principals, or managed identities at a particular scope, such as management group, subscription, resource group, or resource.

To assign roles using the Azure portal, I had to have Microsoft Authorization/role Assignments/write permissions, such as User Access Administrator or Owner.

Then I could assign roles using the Azure portal by following these steps:

Identify the needed scope.

Open the Add role assignment page.

Select the appropriate role.

Select who needs access.

Used compliable languages and tools like C#, ASP.NET

.Net Cross-Platform was used for building the application @ Fisher Investments.

Coordinating with IT Service Management/Development teams to leverage & support capabilities of the ServiceNow CMDB in supporting processes such as, including Incident, Problem, Release Management, Change Management, and Asset Management.

Act as a hands-on mentor and guide for junior team members, leading by example to foster a culture of continuous learning and improvement.

Used Visual Studio 2017 and other Products.

Used IMS (incident management solutions) to streamline incident response communions by using IMS template. That helped us kick off the processes and alert our team with unexpected incident.

For IMS cybersecurity workflows Tines was used. Tines a powerful automation platform that excels our building cybersecurity workflows.

High level testing done by QA but monitored by us (DevOps Team).

The tests included

Static application security testing (SAST) tools and technologies analyze the source code or bytecode from the inside out, helping developers find issues and flaws inside their code. SAST is easily integrated into the software development lifecycle (SDLC) for continuous monitoring.

Interactive Application Security Testing (IAST) is a dynamic security testing method designed to identify vulnerabilities in applications while they are running. Unlike traditional methods, IAST analyzes an application from the inside during runtime, providing real-time feedback on potential security issues.

Software Composition Analysis (SCA) is a security testing methodology that focuses on identifying and managing open-source and third-party components in software applications.

Onsite (San Diego, CA) Dec 2016 – Sep 2017

Sr. Build Engineer/Hands on Manager

Client: Day Break Games LLC. (Sony Corporation Gaming Branch)

Led DevOps transformation initiatives in a high-paced R&D environment

architecting CI/CD pipelines and deployment workflows to enhance automation,

standardization, and developer productivity. Owned key decisions on source control

strategy, environment builds, and toolchain selection to support multi-environment

deployments across AWS and on-prem infrastructure.

Designed and documented comprehensive branching strategies using Perforce Streams to support complex game development lifecycles across Dev, QA, and Production environments.

Applied Release management, for web systems and Content Management Systems (CMS), which was the systematic process of managing and delivering new or updated software components and applications.

Architected and maintained Jenkins-based CI/CD pipelines integrating P4 Streams and parallel build jobs, ensuring efficient multi-platform artifact handling and deployment.

Implemented asset pipeline automation leveraging Jenkins freestyle and multi-job plugins to process binary formats (DDS, SWF) for cross-platform deployment scenarios.

Automated infrastructure processes and enhanced DevOps efficiency by modifying and extending Python and Perl scripts used throughout CI/CD lifecycle.

Deployed builds to AWS environments and streamlined infrastructure configuration for Unix/Linux server estates using PowerShell, shell scripting, and .NET cross-platform tools.

Fostered agile collaboration through daily standups, Jira/Confluence workflows, and R&D planning cycles.

Championed cloud-first and automation-first principles, driving early-stage containerization strategies and fostering a forward-thinking DevOps mindset.

Delivered architectural contributions to infrastructure modernization, laying the groundwork for future Kubernetes integration and low-touch cloud-native workflows.

Onsite (Honolulu, HI) Jul 2016 – Nov 2016

Sr. Software Configuration Engineer/Hands on Manager

Client: Hawaiian Airlines

Team: DevOps

Accomplished/Performed, & maintained the following activities @ NIIT Technologies Hawaiian Airline:

On 07, 2016 – I joined NIIT-Tech and started as a contractor. Started working onsite with one of their clients (Hawaiian Airlines) as Sr. Software Configuration Manager.

Maintain multiple source repository, Source control system (TFS) and branching.

Build & Maintained Applications including SaaS (Software as a Service) applications models.

Automate the build using a cloud version of visual studio to simplify continuous integration of apps for platforms and programming language used. Visual Studio Team System (VSTS) using GIT version 2.10.0. Windows.1.

Visual Studio Team System (VSTS) is an extension of the Microsoft Visual Studio architecture that allows it to encompass development teams, with special roles and tools for software architects, developer specialties, and QA. Visual Studio is a software development environment built on the .NET Framework that is designed for managing projects and development work in a variety of languages including



Contact this candidate