Skip to content

Scripts that I use to automate my hunting. If you want to use, you might need to customize them according to your system. (Frog-man)

Notifications You must be signed in to change notification settings

Frog-man-sec/Autometa

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Autometa: Collection of automation scripts

NOTE

I made these scripts to automate my tasks. So, if you want to use them, I would suggest you to customize these scripts as well as files & folders properly.


recon script

First I will talk about the script recon. This scripts runs some common yet effective recon tools to find subdomains.

For each tool, it stores the results in seperate files.

After that it will combine all the found subdomain and trim out the duplicates.

Task Process

recon script takes sub domain from a text file called scopes.txt. Then it uses the following tools to enumerate subdomains for that domain.
Tools

  • amass
  • subfinder
  • assetfinder
  • crt_org.sh

Meanwhile it stores found subdomains into text files named after the tool that is used to enumerate them.

After completing the enumeration, it marges all the subdomain and put the unique ones into a file called finalDomains.txt

Note: due to the reson that amass takes a lot of time (as I don't have a VPS nor a fast internet connection) I've put timeout function to terminate after certain time. You can remove it if you want

Modification

  • Create a directory called /automate in your /home/<user> directory.

  • Put your script in that directory.

  • Put your urls into a text file called scopes.txt and put it in the same directory as the script.

  • scopes.txt should have clean links. e.g:

    • example.com
    • helloworld.xyz
    • newsite.co
  • Don't put http:// or https:// before the link. e.g:



SubCheck Script

Now as we've done with gathering our subdomains we will move to a part where we will check for possible subdomain takeover.

But if you take a look at the output of recon script, you'll see that, if you provide 2 or more domains, it will create a folder for each of them and store the output in those folder. So, what are you going to do?

  • Go to each folder and run subzy?
  • or Copy all the found sub domains in to one text file and run subzy on it?
  • or Use a script that will do all the hassel for you.

Sub check is for the people who chose option 3. It will cycle through all the folder searching for finalDomains.txt and then test for subdomain takeover.

Just go to the main directory that contains your folders with enumerated subdomains, and run subcheck.sh. If it finds any possible takeover, it will store it at subzy_output.txt.



About

Scripts that I use to automate my hunting. If you want to use, you might need to customize them according to your system. (Frog-man)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages