Tuesday, July 23, 2013

TBH Zone-H Mass Poster




#!/bin/bash 
# TBH Zone-H Mass Poster
#c0ded by Red V!per


notifier=$1
sites=$2
url="http://www.zone-h.org/notify/single"


banner()
{
echo            "                             "
echo -e '\E[32m'"_________ ______             "; tput sgr0
echo -e '\E[32m'"\__   __/(  ___ \ |\     /|  "; tput sgr0
echo -e '\E[32m'"   ) (   | (   ) )| )   ( |  "; tput sgr0
echo -e '\E[32m'"   | |   | (__/ / | (___) |  "; tput sgr0
echo -e '\E[32m'"   | |   |  __ (  |  ___  |  "; tput sgr0
echo -e '\E[32m'"   | |   | (  \ \ | (   ) |  "; tput sgr0
echo -e '\E[32m'"   | |   | )___) )| )   ( |  "; tput sgr0
echo -e '\E[32m'"   )_(   |/ \___/ |/     \|  "; tput sgr0
echo            "                             "
                            
}


Usage()
{
  echo ""
  echo "# ***************************************************************************​****​*********************#"
  echo "# Usage      : TBH Zone-H Mass Poster  <defacer> <sites.lst>                                         *#"
  echo "# Help       : -h && --help : Show This Menu                                                         *#"
  echo "# RunScript  : Give Permision to script and run it !!                                                *#"
  echo "# Example    : ./script.sh www.example.com  ye_bande_khoda sites.lst                                 *#"
  echo "# ***************************************************************************​****​*********************#"
  echo ""
  exit
}


Check_Arguments()
{
  if [ -z "$notifier" ] || [ "$notifier" == "-h" ] || [ "$notifier" == "--help" ]; then
  Usage;
  exit
  fi
}

Mass_Post()
{
for targets in `cat $sites`
do
       curl --user-agent "Mozilla/5.0" --silent --data "defacer=$notifier&domain1=$targets&hackmode=1&reason=1" ${url} &
       sleep 2
done
}

Main()
{
banner;
Check_Arguments;
Mass_Post;
}

Main;

Dowloadn : http://www.mediafire.com/download/j10du919tq25f2d/TBH+Zone-H+Mass+Poster.zip

Saturday, July 20, 2013

Web Exploiter Toolkit



This script makes it easy tasks such as client hacking , web hacking , Wireless hacking ...

Now Working on it ...

https://github.com/RedH4tViper/webexploiter

Red V!per User Domain Extractor


create etc-passwd.txt file Besides User+Domain Extractor script and insert Contents of
/etc/passwd file Within etc-passwd.txt then Give Permision to script and run it


#!/bin/bash
#
# ---------------------------------
# User+Domain Extractor 
# Licence : Linux
# ---------------------------------
#
# Title   : User+Domain Extractor
# Code    : Bash
# Author  : RedH4t.Viper
# Email   : RedH4t.Viper@Gmail.com , RedH4t.Viper@yahoo.com
# Date    : 2013 06 May
# Home    : IrIsT Security Center
# Thanks  : IrIsT ,TBH ,kurdhackteam , 3xp1r3
#
# Gr33tz  : Am!r | C0dex | B3HZ4D | TaK.FaNaR | 0x0ptim0us | Net.W0lf |
# Gr33tz  : Skote_Vahshat| Dj.TiniVini| Mr.XHat | Black King |
# Gr33tz  : E2MAEN | Mr.F@RDIN | M4st3r4N0nY | ICA_r00t | m3hdi |
# Gr33tz  : x3o-1337 | rEd X | No PM  | Gabby | Sukhoi Su-37
# Gr33tz  : ARTA | H-SK33PY | (^_^) | Turk Sever | And All Of IrIsT Memebrz
#------------------------------------------------------------------------------------------#


IP=
F_ARG=$1
page=0
last_page_check=
how_many=1
single_page=


Usage()
{
 echo ""
 echo "# ****************************************************************************************************#"
 echo "# Usage      : User+Domain Extractor <ip>                                                            *#"
 echo "# Help       : -h &amp;&amp; --help : Show This Menu                                                         *#"
 echo "# RunScript  : create etc-passwd.txt file Besides User+Domain Extractor script and insert Contents   *#"
 echo "# RunScript  : of /etc/passwd file Within etc-passwd.txt then Give Permision to script and run it !! *#"
 echo "# ****************************************************************************************************#"
 echo ""
}

get_reverse()
{

# check IP
  if [ `echo "$F_ARG" | egrep  "(([0-9]+\.){3}[0-9]+)|\[[a-f0-9:]+\]"`  ]; then
   IP="$F_ARG"
   else
    IP=`resolveip -s "$F_ARG"`
     if [ "$?" != 0 ]; then
       echo -e "               [-] Error: cannot resolve $F_ARG to an IP "
       exit
     fi
   fi

 url_neighbors="http://www.my-ip-neighbors.com/?domain=$IP"

 wget -q -O  my-ip-neighbors.php "$url_neighbors"

 cat my-ip-neighbors.php | egrep -o '<a -d="" -f4="" cut="" href="http://whois.domaintools.com/%5B%5E" sort="" uniq="">&gt; whois_neighbors.txt


 url_kenh12="http://ip.kenh12.com/Reverse-IP-Lookup-DNS-Domain/$IP.html"
 wget -q -O  my-ip-kenh12.php "$url_kenh12"

 cat my-ip-kenh12.php | egrep -o "gif' alt=[^&gt;]+" | cut -d '=' -f2 | sed '/www./s///g' | sort | uniq &gt;&gt; whois_kenh12.txt


while [ -z "$last_page_check" ] &amp;&amp; [ -n "$how_many" ] &amp;&amp; [ -z "$single_page" ]; do


 url="http://www.bing.com/search?q=ip%3A$IP&amp;go=&amp;qs=n&amp;first=${page}0&amp;FORM=PERE"

 wget -q -O bing.php "$url"

 last_page_check=`egrep -o '<span class="sb_count" id="count">[0-9]+-([0-9]+) of (\1)' bing.php`

 # if no results are found, how_many is empty and the loop will exit
 how_many=`egrep -o '<span class="sb_count" id="count">[^&lt;]+' bing.php |cut -d '&gt;' -f 2|cut -d ' ' -f 1-3`

 # check for a single page of results
 single_page=`egrep -o '<span class="sb_count" id="count">[0-9] results' bing.php`


 cat "bing.php"| egrep -o "</span></span></span></a>

<h3>
<span class="sb_count" id="count"><span class="sb_count" id="count"><span class="sb_count" id="count"><a -d="" -f="" 2="" bing.php="" cut="" href="http://www.blogger.com/%5C">&gt; alldomain_bing.txt
let page=$page+1
 rm -f bing.php
done

cat alldomain_bing.txt | cut -d '/' -f 3 | tr '[:upper:]' '[:lower:]' | sed '/www./s///g' | sort | uniq &gt;&gt; whois_bing.txt
rm alldomain_bing.txt ;


   for line_bing in `cat whois_bing.txt`
     do
       echo "$line_bing" &gt;&gt; reversed.txt
     done

   for line_neighbors in `cat whois_neighbors.txt`
    do
       echo "$line_neighbors" &gt;&gt; reversed.txt
    done
 
   for line_kenh12 in `cat whois_kenh12.txt`
    do
       echo "$line_kenh12" &gt;&gt; reversed.txt
    done

    rm whois_bing.txt;
    rm whois_neighbors.txt;
    rm whois_kenh12.txt;
    rm my-ip-neighbors.php;
    rm my-ip-kenh12.php;

}

get_user()
{
cat etc-passwd.txt | egrep  "/home" | cut -d ':' -f 1 &gt;&gt;user.txt
}

get_user_with_site()
{
for reverse in `cat user.txt`
do
site=`cat reversed.txt | grep "$reverse"`
cat reversed.txt | grep "$reverse" &gt;&gt; /dev/null;check=$?
           if [ $check -eq 0 ]
                then
                echo -e "[+] Found : User: "$reverse" \t Site: $site"
                echo -e "[+] Found : User: "$reverse" \t Site: $site" &gt;&gt;ListUserDomain.txt
           fi
done
rm -rf user.txt
rm -rf reversed.txt
}

main()
{
# check for arguments
if [ -z "$F_ARG" ] || [ "$F_ARG" == "-h" ] || [ "$F_ARG" == "--help" ]; then
Usage;
exit
fi

get_reverse;
get_user;
get_user_with_site;
}

main;

Download: http://www.mediafire.com/download/ismd5f47n7hs7my/User_Domain_Extractor.sh 

Red V!per Sub Domain Scanner



#!/bin/bash
# WebExploiter Sub Domain Scanner
# Coded By : Red H4t V!per

page=0
how_many=1
URL=$1
domain= 
single_page=
last_page_check=

banner()
{
echo "           _         _                       _                                              "
echo "          | |       | |                     (_)                                             "
echo " ___ _   _| |__   __| | ___  _ __ ___   __ _ _ _ __    ___  ___ __ _ _ __  _ __   ___ _ __  "
echo "/ __| | | | '_ \ / _\ |/ _ \| '_ \ _ \ / _\ | | '_ \  / __|/ __/ _\ | '_ \| '_ \ / _ \ '__| " 
echo "\__ \ |_| | |_) | (_| | (_) | | | | | | (_| | | | | | \__ \ (_| (_| | | | | | | |  __/ |    "
echo "|___/\__,_|_.__/ \__,_|\___/|_| |_| |_|\__,_|_|_| |_| |___/\___\__,_|_| |_|_| |_|\___|_|    "
echo "                                                                                            "

}
Usage()
{
echo ""
echo "# ***************************************************************************​****​*********************#"
echo "# Usage      : WebExploiter Sub Domain Scanner <domin>                                               *#"
echo "# Help       : -h && --help : Show This Menu                                                         *#"
echo "# RunScript  : Give Permision to script and run it !!                                                *#"
echo "# ***************************************************************************​****​*********************#"
echo ""
}

Check_Arguments()
{
if [ -z "$URL" ] || [ "$URL" == "-h" ] || [ "$URL" == "--help" ]; then
Usage;
exit
fi
}

Sub_Domain()
{

rm -f alldomain_bing.txt;
rm -f sub.txt;
rm -f sub2.txt;


if [ `echo "$URL" | egrep  "http://"` ] || [ `echo "$URL" | egrep  "Http://"` ]; 
then
echo -e "\e[1;31m[-] please insert domain with out http:// \e[0m" 
 exit 1
fi

if [ `echo "$URL" | egrep  "www."`  ]; 
then
domain=`echo "$URL" |sed '/www./s///g' `
else 
domain=$URL
fi


echo -e "\e[1;31m[*] SubDomain Urls Will be Saved  at subdomain.txt Searching Plz W8...\e[0m" 


while [ -z "$last_page_check" ] && [ -n "$how_many" ] && [ -z "$single_page" ]; do


url="http://www.bing.com/search?q=%27$domain%27&qs=n&pq=%27$domain%27&sc=0-0&sp=-1&sk=&first=${page}0&FORM=PERE"

wget -q -O sub_domain_bing.php "$url"

 last_page_check=`egrep -o '<span class="sb_count" id="count">[0-9]+-([0-9]+) of (\1)' sub_domain_bing.php`

 how_many=`egrep -o '<span class="sb_count" id="count">[^<]+' sub_domain_bing.php | cut -d '>' -f 2|cut -d ' ' -f 1-3`

 single_page=`egrep -o '<span class="sb_count" id="count">[0-9] results' sub_domain_bing.php `


 cat "sub_domain_bing.php" | egrep -o "<h3><a href=\"[^\"]+" sub_domain_bing.php | cut -d '"' -f 2 >> alldomain_bing.txt
 rm -f sub_domain_bing.php
 let page=$page+1 
 done

cat alldomain_bing.txt | cut -d '/' -f 3 | tr '[:upper:]' '[:lower:]' | sed '/www./s///g' | sort | uniq >> sub.txt

for var in `cat sub.txt` 
do 
echo "$var" | grep "$domain" >> /dev/null;check=$? 
if [ $check -eq 0 ] 
then 
 echo "$var" >> sub2.txt
fi 
done

cat sub2.txt | sort | uniq > subdomain.txt


found_N=`wc -l subdomain.txt | sed 's/subdomain.txt//'` 
echo -e "\e[1;34m[+] Found $found_N SubDomain :) \e[0m" 


for catvar in `cat subdomain.txt`
do
echo -e "\e[1;32m[*] $catvar \e[0m"   
done 

rm -f alldomain_bing.txt;
rm -f sub.txt;
rm -f sub2.txt;
}

main ()
{
banner;
Check_Arguments;
Sub_Domain;
}

main ;

Download : http://www.mediafire.com/download/638lw6ao7zjw7c4/red-viper-sub-dmoain-scannner.sh

Red V!per Server joomla Finder

 

#!/bin/bash
#
# ---------------------------------
# Server joomla Finder  
# Licence : Linux
# ---------------------------------
#
# Title    : Server Joomla Finder 
# Code     : Bash
# Author   : RedH4t.Viper
# Email    : RedH4t.Viper@Gmail.com , RedH4t.Viper@yahoo.com
# Released : 2013 04 June
# Thanks   : IrIsT ,TBH ,kurdhackteam , 3xp1r3 , thecrowscrew
#
# Gr33tz   : Am!r | C0dex | B3HZ4D | TaK.FaNaR | 0x0ptim0us | Skote_Vahshat | 
# Gr33tz   : Net.W0lf | Dj.TiniVini| Mr.XHat | Black King | Devil | 
# Gr33tz   : E2MAEN | (^_^) | M4st3r4N0nY |Turk Sever | dr.koderz | V30sharp
# Gr33tz   : ARTA | Mr.Zer0 | Sajjad13and11 | Silent | Smartprogrammer |  
# Gr33tz   : x3o-1337 | rEd X | No PM | Gabby | Sukhoi Su-37 | IR Anonymous |
# Gr33tz   : Megatron | Zer0 | sole sad | Medrik | F@rid | And All Of IrIsT Memebrz |
#------------------------------------------------------------------------------------------#

page=0
how_many=1
IP_SERVER=$1
single_page=
last_page_check=

banner()
{
echo "     _                       _          __ _           _              "
echo "    (_)                     | |        / _(_)         | |             "
echo "     _  ___   ___  _ __ ___ | | __ _  | |_ _ _ __   __| | ___ _ __    "
echo "    | |/ _ \ / _ \| '_ \ _ \| |/ _\ | |  _| | \_ \ / _\ |/ _ \  __|   " 
echo "    | | (_) | (_) | | | | | | | (_| | | | | | | | | (_| |  __/ |      "
echo "    | |\___/ \___/|_| |_| |_|_|\__,_| |_| |_|_| |_|\__,_|\___|_|      "
echo "   _/ |                                                               "
echo "  |__/                                                                "   
echo "                                                                      "
}

Usage()
{
echo ""
echo "# ***************************************************************************??****?*********************#"
echo "# Usage      : Server Joomla Finder  <IP/Domain>                                                     *#"
echo "# Help       : -h && --help : Show This Menu                                                         *#"
echo "# RunScript  : Give Permision to script and run it !!                                                *#"
echo "# ***************************************************************************??****?*********************#"
echo ""
}

Check_Arguments()
{
if [ -z "$IP_SERVER" ] || [ "$IP_SERVER" == "-h" ] || [ "$IP_SERVER" == "--help" ]; then
Usage;
exit
fi
}

Searching_Jce()
{
 rm -rf domains.txt
 rm -rf alldomain_bing.txt
 rm -rf IndexDomain.txt

  if [ `echo "$IP_SERVER" | egrep  "(([0-9]+\.){3}[0-9]+)|\[[a-f0-9:]+\]"`  ]; then
    IP="$IP_SERVER"
   else
     IP=`resolveip -s "$IP_SERVER"`
      if [ "$?" != 0 ]; then
       echo -e "[-] Error: cannot resolve $IP_SERVER to an IP " 
        Check_Back
      fi
    fi

echo -e "\e[1;35m[*] Finded Joomla Web Sites Will be Save at finded.txt \e[0m"
echo -e "\e[1;35m[*] Searching Joomla Web Sites on $IP Plz W8 \e[0m"


touch  alldomain_bing.txt;

 while [ -z "$last_page_check" ] && [ -n "$how_many" ] && [ -z "$single_page" ]; do

 url="http://www.bing.com/search?q=ip%3a$IP+%27index.php?option=com_%27&qs=n&pq=ip%3a$IP+%27index.php?option=com_%27&sc=8-26&sp=-1&sk=&first=${page}1&FORM=PERE"


 wget -q -O domain_bing.php "$url"


 last_page_check=`egrep -o '<span class="sb_count" id="count">[0-9]+-([0-9]+) of (\1)' domain_bing.php`

# if no results are found, how_many is empty and the loop will exit
 how_many=`egrep -o '<span class="sb_count" id="count">[^<]+' domain_bing.php | cut -d '>' -f 2|cut -d ' ' -f 1-3`

#  check for a single page of results
 single_page=`egrep -o '<span class="sb_count" id="count">[0-9] results' domain_bing.php `


 cat domain_bing.php | egrep -o "<h3><a href=\"[^\"]+" domain_bing.php | cut -d '"' -f 2 >> alldomain_bing.txt
 rm -f domain_bing.php
 let page=$page+1 
 done


 cat alldomain_bing.txt | awk '{gsub("http://","")}1' | awk '{gsub("https://","")}1' | sed '/www./s///g' | tr '[:upper:]' '[:lower:]' | sort | uniq >> domains.txt


  for domain in `cat domains.txt`
   do
    echo "$domain" | grep "com_" >> /dev/null;check=$?
               if [ $check -eq 0 ]
                    then
                   echo "$domain" >>IndexDomain.txt
               fi
    done

cat IndexDomain.txt | cut -d '?' -f 1 | awk '{gsub("index.php","")}1' | sort | uniq >> finded.txt

found_N=`wc -l finded.txt | sed 's/finded.txt//'` 
echo -e "\e[1;34m[+] Found $found_N :) \e[0m" 


for joom in `cat finded.txt`
do
echo -e "\e[1;32m[*] $joom \e[0m"   
done 

rm -rf domains.txt
rm -rf alldomain_bing.txt
rm -rf IndexDomain.txt
}

main()
{
banner ;
Check_Arguments;
Searching_Jce;
}

main;   

Download : http://www.mediafire.com/download/mtpy84d7lp0byol/red-viper-joomla-finder.sh

Red V!per Email Collector Bash Script

#!/bin/bash
#
# ---------------------------------
# TBH && IrIsT EMail Collector 
# Licence : Linux
# ---------------------------------
#
# Title   : TBH && IrIsT EMail Collector
# Code    : Bash
# Author  : RedH4t.Viper
# Email   : RedH4t.Viper@Gmail.com , RedH4t.Viper@yahoo.com
# Date    : 2013 09 May
# Thanks  : IrIsT ,TBH ,kurdhackteam , 3xp1r3
#
# Gr33tz  : Am!r | C0dex | B3HZ4D | TaK.FaNaR | 0x0ptim0us | Net.W0lf |
# Gr33tz  : Skote_Vahshat| Dj.TiniVini| Mr.XHat | Black King |
# Gr33tz  : E2MAEN | Mr.F@RDIN | M4st3r4N0nY | ICA_r00t | m3hdi |
# Gr33tz  : x3o-1337 | rEd X | No PM  | Gabby | Sukhoi Su-37
# Gr33tz  : ARTA | H-SK33PY | (^_^) | Turk Sever | And All Of IrIsT Memebrz
#------------------------------------------------------------------------------------------#
page=0
last_page_check=
how_many=1
single_page=
Domain=$1

Usage()
{
 echo ""
 echo "# ****************************************************************************************************#"
 echo "# Usage      : TBH EMail Collector <domain>                                                          *#"
 echo "# Help       : -h && --help : Show This Menu                                                         *#"
 echo "# RunScript  : Give Permision to script and run it !!                                                *#"
 echo "# ****************************************************************************************************#"
 echo ""
}

# check for arguments
if [ -z "$Domain" ] || [ "$Domain" == "-h" ] || [ "$Domain" == "--help" ]; then
Usage;
exit
fi

 while [ -z "$last_page_check" ] && [ -n "$how_many" ] && [ -z "$single_page" ]; do

 url="http://www.bing.com/search?q=%27%40$Domain%27&qs=n&pq=%27%40$Domain%27&sc=0-0&sp=-1&sk=&first=${page}1&FORM=PERE7"

 wget -q -O page.php "$url"


 last_page_check=`egrep -o '<span class="sb_count" id="count">[0-9]+-([0-9]+) of (\1)' page.php`

# if no results are found, how_many is empty and the loop will exit
 how_many=`egrep -o '<span class="sb_count" id="count">[^<]+' page.php | cut -d '>' -f 2|cut -d ' ' -f 1-3`

#  check for a single page of results
 single_page=`egrep -o '<span class="sb_count" id="count">[0-9] results' page.php`

 cat page.php | egrep -o "[-A-Za-z0-9_.]+<strong>@$Domain</strong>" | tr '[:upper:]' '[:lower:]' | sort | uniq>>mails.lst

 rm -f page.php
 let page=$page+1
 done

cat mails.lst | cut -d '<' -f 1 | tr '[:upper:]' '[:lower:]' | sort | uniq >> nmail.lst

found_N=`wc -l nmail.lst | sed 's/nmail.lst//'`
echo -e "\e[1;34m [+] $found_N Email Collocted :) \e[0m"

for ec in `cat nmail.lst`
do
echo "[*] $ec@$Domain" >> Collected.lst  
done


for ec in `cat nmail.lst`
do
echo -e "\e[1;32m[*] $ec@$Domain \e[0m"  
done

rm -rf mails.lst
rm -rf nmail.lst  


Download : http://www.mediafire.com/view/9962o54dqx820ow/redviper-email-collector.sh

Friday, July 19, 2013

How to Attack on Remote PC With Applications Vulnerabilities





Back Track Tutorials 

Article Name : How to Attack on Remote PC With Applications Vulnerabilities
language : English
Page # : 56

Download : www.mediafire.com/?809b6bu3txi2v1a