Research Button

URL scanning for new age security measures

It’s the URL, stupid (me)!

Consider a scenario. You are in a miserable situation where you accidentally clicked on some phishing link or scam URL. A long time ago, when the web was safe, and viruses, trojans, and worms were transmitted only by EXE or BIN files, we could rest assured that the virus scanner protected us.

Now, the web is the purveyor of all things good and evil. Smartphones have become the norm rather than the exception. The individual security measures (windows, IOS, etc.) are only as useful as there latest update, and maintaining the amount of tech we each possess up to date is difficult. But the most common denominator is the URL centric web. All devices have Internet access and thus are vulnerable to the latest threats.

Think of the email before spam abuse. There used to be open relays everywhere, and anyone could send emails using a 10 line shell script using SMTP command verbs. Today that is impossible since email abuse has turned people away from everyday email. Even when you need to use email for work, most of one’s inbox is someone trying to sell you something or market something. Every piece of traffic that humans originated on the Internet has a URL, so let’s look at new-age security measures to help protect you.

Let’s safely examine a URL

We can’t trust all the URLs friends and family send over as many social media sites, emails, and whatnot have been compromised. Most likely, each of us has received a malicious URL coming from close friends’ or family members’ accounts. What choices do we have?

(a) ignore and appear rude
(b) pray and click
(c) investigate further before clicking

For the diligent ones amongst us, what is the best way to investigate further one may ask?

One option is a background scanner that looks at each website’s health the URL points to without eyeballing it manually. Wouldn’t that be nice? Nowadays, it is not the OS on your desktop but the browser web access that dictates your computer health and network resource usage. By scanning a URL using back-end tools like UrlScan, we can get a lot of information about how the website looks and works, including its dynamic content, video, images, etc.

There are several tools but let’s start with UrlScan. UrlScan is a Microsoft tool that analyzes the text content and the dynamic javascript and DOM structure to filter out unwanted websites. The idea is that if a website URL is unsafe, a 404 (not found error) is returned, thereby making the click work safely even with automated software-based clicks and accesses.

Another such tool:, cans and analyses websites. When a URL goes to, an automated process browses it (like a regular user) and records the activity on that page. This process discovers the domains and IPs contacted, the resources (JavaScript, CSS, etc.) requested from those domains, and additional information about the page itself. will take a screenshot of the page, record the DOM content, JavaScript global variables, cookies created by the page, and many other observations. If the site targets the users of a familiar brand tracked by, it is highlighted as potentially malicious in the scan results.

I have seen plenty of open source tools that use UrlScan for various purposes, even to break down an HTML file into links inside the mutt email client on Linux. UrlScan operates using a synchronous call method in which you wait for results when you invoke the scan. But it is commonly used by submitting a URL and then polling for scan results later. Due to the high traffic and owing to the service being free, UrlScan requires you to obtain a free API key for the scan service and is hassle-free in my experience. But to use the service, we must implement exponential backoff and play like a good citizen and not abuse their server resources.

Here are some examples to get you started.

I am using urlscanio, a github project ( ), for a Python CLI for First you export the API key like this:

$ export URLSCAN_API_KEY=<key>

$ urlscanio -i

Scan report URL:
Screenshot download location: screenshots/dc7e62d6-8738-4fd2-bddf-6b9c42b8ff2a.png
DOM download location: doms/dc7e62d6-8738-4fd2-bddf-6b9c42b8ff2a.txt

You also have the --submit and --retrieve commands for the offline operation and batch requests.

If you look at the screenshots you see this.

Screen Shot 2020-10-01 at 2.18.17 PM

And the scan URL shows this:
Screen Shot 2020-10-01 at 2.18.30 PM

UrlScan is a method to vet website URLs and is making a splash in the security marketplace. But it has a lot of potential as UrlScan is free for use for limited query rates, can be inserted into all sorts of REST API services and scanning algorithms.

As opposed to command-line based and browser-based scanners, this back-end based service, which uses a Google Chrome headless method, is beneficial and powerful. is a community website with some sponsors, and Microsoft has released a UrlScan tool to be used with IIS servers. Also, you have several open-source tools on Github with the ability to scan URLs. The goal here is, instead of manually inspecting a website, its dynamic and static content, the web crawling is done by a back-end tool, and you get an idea about how safe the URL is. The URL itself could point to a video, an image file, audio, or a website. Several such scanners exist, with each of them offering a different accuracy.

VirusTotal is a Spanish company that came up with a similar service long ago. Since then, Google acquired them, and it has been a service ever-growing in popularity. I will cover it further in a separate article.


Conclusion and evolution from here

The URL scanning technique is catching on in today’s web and is built into many software pre-processing pipelines. But this also means that just like we get false positives with spam control, URL scanning is also not a perfect art, so you might have to intervene from time to time and ferret out the decision-making process to make it work correctly.

There are more tools like the three mentioned above, and they are an excellent way to guard against phishing and abuse from URL based cloud services and other potentially harmful websites. Most unsuspecting users lose money or information online through dangerous website links that dupe people into installing trojan horses, worms, or launching attacks into your network.

Implementing a pipeline that employs this ground zero filterings is up to you. Still, as an ab initio method to filter harmful URL accesses, this technique like UrlScan is undoubtedly valuable. But to benefit from it, you must build other tools into the mix and manually inspect the background scan screenshots and such. This is an open-source gratis and done in the background, which opens up a pandora’s box of possibilities from a programmer’s perspective.


Is there a non-programmer way?

Writing URL scanning Python pipelines is excellent for those with the skills, time, and resources, but their security analyst resources are so scarce that it is infeasible for most organizations. So what do we do? Bit of a shameless plug, but Query.AI ( ), where I work, integrates URL scanning tools and provides an easy and efficient way to safely examine URLs without needing any Python or other scripting methods.

Did you enjoy this content? Follow our linkedin page!

Looking for similar content?