bookmark_borderUsing Pipelines to Invalidate AWS CloudFront Cache that Pointing to AWS S3 bucket

Steps:

  • Add Repository Variables for Distribution IDs
  • Add Repository Variables for AWS keys as well
  • Add invalidate step in your pipeline
    e.g.
- step:
          name: ">> Invalidate AWS CloudFront (by: Khaled alam)"
          script:
          - pipe: atlassian/aws-cloudfront-invalidate:0.4.1
            variables:
              AWS_ACCESS_KEY_ID: $AWS_ACCESS_KEY_ID
              AWS_SECRET_ACCESS_KEY: $AWS_SECRET_ACCESS_KEY
              AWS_DEFAULT_REGION: $AWS_DEFAULT_REGION
              DISTRIBUTION_ID: $TEST_CLOUDFRONT_DISTRIBUTION_ID
              # PATHS: <string> # Optional
              DEBUG: "true" # Optional

Note: in case you want to invalidate specific files or paths, you should define that in PATHS: <string>

bookmark_borderAPI Multiprocessing

Motivation

In this article, I am going to show you how we can improve the computing power of simple API script from total overall (6 minutes and 17 seconds) to (1 minute 14 seconds)

The Idea

I will share with you one of my simple favourite technique that I prefer to use especially when I work on data science tasks such as data visualization, data analysis, code optimization, and big data processing.

Processing a task in a sequential way may take a long time especially when we are talking about a huge amount of data(eg. big inputs)

This technique takes advantage of parallelization capabilities in order to reduce the processing time.

The idea is to divide the data into chunks so that each engine takes care of classifying the entries in their corresponding chunks. Once performed, each engine reads, writes and processes its chunks, each chunk be processed in the same amount of time.

Example

The example I choose to use for this article is Genderize names that consist of 2 alphabetic characters.

Output Analysis Chart

Explanation

Clone GitHub Repo and follow instructions in Usage section.

Let’s generate all alphabet names that consist of 2 characters(to make the testing process easy)

we can use some Linux Kali penetration testing tool[1] such as crunch
$ crunch 2 2 > names.txt
so we generate all possible alphabet names with length 2 (676 lines)

then let’s create directories which are needed for splitting process
$ mkdir subs/ subs/inputs subs/outputs subs/outputs/parts subs/outputs/all

now we can split out input data, there are many ways to do that but I prefer to use Unix split command [2]
$ split -l 100 -d names.txt ./subs/inputs/
so we split names.txt file into small files, each file consists of 100 lines

now let’s run all processes: ./init.bash
after finish use merger.py script to merge all outputs.
merging process separated to avoid conflicts behaviours and sorting-save.


The Project on GitHub:
https://github.com/khaledalam/api-multiprocessing

An application uses this technique:
Hiring-Related Email(https://github.com/khaledalam/amazon-jobs)

Interesting related ideas:
Parallelizing using GPUs
– MapReduce (https://en.wikipedia.org/wiki/MapReduce)

[1] https://tools.kali.org/password-attacks/crunch
[2] https://en.wikipedia.org/wiki/Split_(Unix)

bookmark_borderAmazon Job Hiring Related Email

Amazon-Jobs

Amazon Jobs Apps that help the user to check if some email relates to any AMAZONIAN hrs, recruiters, interviewees, interviewers, talents, candidates and so on..

  • – check hiring-related emails,
  • – get job post recruiter details,
  • – get more statistics about job posts,
  • – search jobs globally with more deep filters

Preview:




Mobile App:


Web App:
https://khaledalam.net/amazon

GitHub:



Screens:


bookmark_border2 SDE Amazon Interviews Invitations in 1 week

2 SDE Amazon interviews invitations in 1 week, a new experience!


Germany


Spain


Germany


I used to get rejections from Amazon at CV monitoring stage :’( but I never give up! I also used to feel this low energy after finish contests ex. interesting codeforces rounds, although sometimes in contests I solve problems that are much harder than multi-international companies interviews questions but I have to admit that after these 2 interviews my battery is not low as usual it literally dies instead :’)


Let’s analyze briefly:

  • Total time [1.5 : 2.5] hours
  • Total questions topics-based: DP, BT, Recursion, String Manipulation, basic math, build and sort complex ds.
  • A session for algorithms & data-structures coding questions.
  • A session for open-ended questions to discuss your solutions and complexity.
  • A session for reasoning questions(very tricky).
  • A session for code debugging ability(not hard).
  • A session for working-style questions(focused on soft skills + psychological dimensions).
  • A session for a survey.

Notes:

  • Tricky corner test cases.
  • DS coding questions are annoying and need to start with wisely choices and smart ideas from the beginning.
  • Open-ended questions have a very short time, need to think and organize your answer in your mind before the session or during the coding session.
  • Overall Div2 ~ D level can nail it.

For sure there are other hundreds of questions topics, sessions, and styles but this was my own experience!

If you are still an undergraduate, my advice: “problem-solving” & “practice”

My greetings < 3