Working with the TechNow lab for the PA-215: Palo Alto Networks Firewall Essentials FastTrack course has been nothing less than a techie's idea of fun.  When students come in we are immediatly configuring the Cisco 3750 switches for access ports, VLANS, and trunks.  We then cable the switch to the Palo Alto Networks Firewall.  Each student gets their own Palo Alto Firewall Pod of hardware and software.  What we find as fun is the VLAN environment, with an array of virtual machines hosted on an ESXi server that can really exercise the abilities of the Palo Alto Firewall.  The DMZ VLAN hosts virtual machines that support enterprise services and also potentialy vulnerable web services.  The Trust VLAN has Windows and Linux clients.  The UnTrust VLAN has Web services and a VM of Kali. The hardware Firewall is additionally connected to a Management VLAN.  All those VLANs are trunked into an ESXi server where the student also has a VM-Series Palo Alto Networks Firewall for High Availability.  

After configuring all the trunking, VLANs, and network interfaces we learn about the firewall and configure it for the lab environment.  Using Metasploitable and Kali/Metasploit nefarious penetration attempts are executed.  Using packet captures, custom APP-ID's  and custom signatures are generated.  Custom logging and reporting are created to similate and enterprise and assist the desired Incident Response.  It is always fun in a training environment to learn all about the controls available in a product, even though specific controls may not be used in the operational environment.  In the end we have a good understanding of the Palo Alto Networks Firewall.

 

Course Overview:

In this course, the students will design various data platform technologies into solutions that are in line with business and technical requirements. This can include on-premises, cloud, and hybrid data scenarios which incorporate relational, NoSQL, or Data Warehouse data. They will also learn how to design process architectures using a range of technologies for both streaming and batch data. The students will also explore how to design data security, including data access, data policies, and standards. They will also design Azure data solutions, which includes the optimization, availability, and disaster recovery of big data, batch processing, and streaming data solutions.

TechNow has worked worldwide enterprise infrastructures for over 20 years and has developed demos and labs to exemplify the techniques required to demonstrate cloud technologies and to effectively manage security in the cloud environment.

Attendees to DP-201: Designing an Azure Data Solution will receive TechNow approved course materials and expert instruction.

Date/Locations:

No Events

Course Duration: 3 days

Course Outline:

  • Data Platform Architecture Considerations
  • Azure Batch Processing Reference Architectures
  • Azure Real-Time Reference Architectures
  • Data Platform Security Design Considerations
  • Designing for Resiliency and Scale
  • Design for Efficiency and Operations

Prerequisites :

      • In addition to their professional experience, students who take this training should have technical knowledge equivalent to the following courses:
      • AZ-900: Microsoft Azure Fundamentals
      • DP-200: Implementing an Azure Data Solution

Comments

Latest comments from students


 

Liked the class?  Then let everyone know!

 

Course Overview: 

This course details how a well-educated Product Owner and a knowledgeable Scrum Master can work together to deliver a successful product with Agile principles using SCRUM approach.  Designed specifically for students who want to learn the mechanics of an Agile / Scrum team being led by a Product Owner and a Scrum Master.

This course offers a firm grasp of Agile principles as they relate to new product development.  If you need to learn what is required to lead or participate in an Agile effort using the SCRUM approach within your organization, this course covers the Agile adaptive life cycle framework and everything in between.

Attendees to PM-232: Product Owner and Scrum Master Roles in AGILE using SCRUM will receive TechNow approved course materials and expert instruction.

Dates/Locations:

No Events

Duration: 2 Days

Course Objectives: At the conclusion of this course, students will be able to:

  • Understand the Scrum Flow, the core components of the Scrum framework
  • Understand the principles of empirical process control
  • Understand the scope of the Product Owner role in detail
  • Understand the scope of the Scrum Master role at a high level
  • Understand the scope of the Scrum team roles and why there is no project manager
  • Understand how the Scrum Master measures team velocity
  • Understand the importance of having the product vision as an overarching goal galvanizing the entire Scrum team
  • Understand the relationship between the vision and the product roadmap
  • Understand the different estimation levels in Scrum
  • Understand what the Product Backlog is and what it is not
  • Understand Product Backlog grooming
  • Understand that Scrum planning is adaptive, iterative, incremental, and collaborative

 

Target Student:

  • Designed specifically for Agile project team members, product owners, project leaders and senior managers or anyone wanting to understand the Agile Framework.

 

Comments

Latest comments from students


Liked the class?  Then let everyone know!

 

Course Overview:

Through hands-on labs, you will learn to automate system administration tasks on managed hosts with Ansible, learn how to write Ansible playbooks to standardize task execution, and manage encryption for Ansible with Ansible Vault. This course will also teach you how to deploy and use Red Hat® Ansible Tower to centrally manage existing Ansible projects, playbooks, and roles; perform basic maintenance and administration of the Ansible Tower installation; and configure users and teams and use them to control access to systems, projects, and other resources through role-based access controls. You will learn to use Ansible Tower’s visual dashboard to launch, control, and monitor Ansible jobs; use the Ansible Tower application programming interface (API) to launch jobs from existing templates; automatically schedule Ansible jobs; and dynamically update host inventories.

Course Objectives:

  • Install and troubleshoot Ansible on central nodes and managed hosts
  • Automate administration tasks with Ansible playbooks and ad hoc commands
  • Write effective Ansible playbooks
  • Protect sensitive data used by tasks with Ansible Vault.
  • Install and configure Ansible Tower for enterprise Ansible management
  • Use Ansible Tower to control access to inventories and machine credentials by users and teams
  • Create job templates in Ansible Tower to standardize playbook execution.
  • Centrally launch playbooks and monitor and review job results with Ansible Tower

 

Course Outline:

  • Introduce Ansible
  • Deploy Ansible
  • Implement playbooks
  • Manage variables and inclusions
  • Implement task control
  • Implement Jinja2 templates
  • Implement roles
  • Configure complex playbooks
  • Implement Ansible Vault
  • Troubleshoot Ansible
  • Install Ansible Tower and describe Ansible Tower’s architecture
  • Create users and teams for role-based access control
  • Create and manage inventories and credentials
  • Manage projects for provisioning with Ansible Tower
  • Construct advanced job workflows
  • Update inventories dynamically and compare inventory members
  • Maintenance and administration of Ansible Tower

 
Dates/Locations:

No Events

Duration: 5 Days

Prerequisites:

  • Become a Red Hat Certified System Administrator, or demonstrate equivalent experience

Target Audience:

This course is designed for Linux system administrators, cloud administrators, and network administrators needing to automate configuration management, application deployment, and intraservice orchestration at an enterprise scale.

 

Course Overview:

In this course, the students will implement various data platform technologies into solutions that are in-line with business and technical requirements, including on-premises, cloud, and hybrid data scenarios incorporating both relational and NoSQL data. They will also learn how to process data using a range of technologies and languages for both streaming and batch data.

The students will also explore how to implement data security, including authentication, authorization, data policies, and standards. They will also define and implement data solution monitoring for both the data storage and data processing activities. Finally, they will manage and troubleshoot Azure data solutions which includes the optimization and disaster recovery of big data, batch processing, and streaming data solutions.

TechNow has worked worldwide enterprise infrastructures for over 20 years and has developed demos and labs to exemplify the techniques required to demonstrate cloud technologies and to effectively manage security in the cloud environment.

Attendees to DP-200: Implementing an Azure Data Solution will receive TechNow approved course materials and expert instruction.

Date/Locations:

No Events

Course Duration: 4 days

Course Outline:

  • Azure for the Data Engineer
  • Working with Data Storage
  • Enabling Team Based Data Science with Azure Databricks
  • Building Globally Distributed Databases with Cosmos DB
  • Working with Relational Data Stores in the Cloud
  • Performing Real-Time Analytics with Stream Analytics
  • Orchestrating Data Movement with Azure Data Factory
  • Securing Azure Data Platforms
  • Monitoring and Troubleshooting Data Storage and Processing

Prerequisites :

      • In addition to their professional experience, students who take this training should have technical knowledge equivalent to the following courses:
      • AZ-900: Microsoft Azure Fundamentals

Comments

Latest comments from students


 

Liked the class?  Then let everyone know!