Hitchhiker’s Guide To Swe At Amazon

3.1 million servers

Data centers

Virtualization: Taking a real machine, and converting it something like 128 virtual machines (for more efficiency)

  • Bare-metal is a name for real machines, as opposed to virtual machines

Hostclass

  • Primary owner
  • Financial owner
  • Operators

Permissions/POSIX group

  • My group, I can give permissions to everyone in my group if they don’t have it yet

Teams:

  • People who report to Lucy, automatically updates and makes a POSIX group, replaced posix Group in a sense, don’t have to worry about it now.

Networks

PROD

  • network that all public facing stuff is deployed do

CORP

  • internal stuff, printers, routers, IP addresses etc
  • also where we put our fulfillment centers
  • Integration tests/servers are probably running in the corp network

Load balancer

  • Group a series of hosts as a series of hosts, depending on how much processing they need
  • also called VIPs (virtual IP address), as they automatically determine where the IP should go behind the scenes (like virtual memory)

Regions

  • aggregate of data centers into a geographical region

Network Regionalization

  • Firewall that blocks stuff going from different parts of your network
  • opens up a lot of scaling growth to use IP addresses in more ways

Snowforting

  • Prod is split into AWS segment and retail segment
  • Makes sure that internal amazon and consumer facing network isn’t on the same network

SOA (Service-Oriented Architecture)

  • Instead of big monoliths, you have lots and lots of small services, each with an interface and individual way to interact with them.
  • 2pt teams (small teams that have ownership of their software)

Microservices

  • breaking things up into smaller and smaller pieces
  • cupcakes - all independent
  • 10s of thousands of microservices

Every time you build a new service, there’s work you have to do again and again and again, so amazon did something about this:

Coral

  • most popular service framework at amazon
  • scheduling requests between services
    • making sure that they’re not overloaded

Brazil

  • build process, from gitfarm gets built into softwares
  • then gets turned into packages (bundles that are deployable as a unit)
    • these can get deployed onto machines or used in other packages

Apollo 202007061518

  • Apollo takes packages and puts them onto machines
  • Used to be Houston, but someone made a change that messed up things for everyone, and they realized that things weren’t very scalable.
  • environment = package + hostclasses
  • stages are host classes that are part of host class definition (devo, alpha, gamma etc)

Pipelines

  • orchestrates development from range of different deployments and environments

PMET (Time series metrics)

  • what was the measurement over a certain time period?
  • measure response times for customers, how much cpu /memory I’m using, etc

IGRAPH / monitor portal

  • can see metrics visualized

AWS

Everything else was kind of old, now aws is the new, and people are moving towards native aws for everything

Conduit and Isengard

  • creating and securing aws accounts

EC2

Autoscaling

  • adds hosts/instances to your fleet to meet traffic demand

Autoscaling group

  • Collection of instances, you can give it rules about how many you want to start with

Availability zone

  • know that this might be available to different zones, changes of 2 zones being impacted by the same event is unlikely

VPC

  • virtual private cloud
  • this is my private virtual network within AWS

MAWS (Move to AWS)

  • moved gurupa (rendering platform on website)
  • direct connect - connecting between on premise network and AWS
  • Apollo can’t deploy to EC2 instances, so you need Apollo Cloud Control (ACC)

uid: 202007061100 tags: #amazon


Date
February 22, 2023