The Wayback Machine - https://web.archive.org/web/20160201183959/http://www.explorehacking.com/search/label/google
Showing posts with label google. Show all posts
Showing posts with label google. Show all posts

Google Dorks | Using Google efficiently

Do you know how to use google ? Sounds a silly question because even children can easily use google search engine.But the thing is that how efficiently we can make use of google. Google dorks are nothing but simple search operators that are used to refine our search. okay lets suppose that you wanna search for ebooks on topic 'networking' , our obvious search queries would be like this "Networking ebooks", "free ebooks for networking" etc,we keep going into the websites, clicking on link after links and then get proper downling links. Now lets do the same search in a different way , type on google "ext:pdf networking" (without quotes)


and see what you get in results,google returned direct downloading links of ebooks on networking that is files with extension pdf.
I hope you have got an idea of google dorks or google search operators.

List of basic search oprators,their description and examples.

1. site - It returns the websites of specified domains .

Example- site:explorehacking.com will return the links of webpages of  domain explorehacking.com. site:explorehacking.com phishing will return all the webpages of domain explorehacking.com which contain word 'phishing'.
Suppose you want google to return only government websites,
you can use site:gov or say pakistani websites use site:pk

2.intitle and allintitle - It restricts the results to pages whose title contain specified word/phrase.

 examples-
intitle:admin will return only those pages whose title contains word 'admin'.
intitle:admin login will return only those pages whose title contains word 'admin' and word 'login' can be in anywhere in page.
allintitle:admin login will return those pages whose title contains both words admin and login. This is simply equivalent to intitle:admin intitle:login

3.inurl- as clear from name, it restricts the results to sites whose URL's contains specified phrase.

examples-
inurl:adminlogin will return only those pages whose URL contains  'adminlogin'.
 Like allintitle, you can similarly use allinurl. I dont feel there is need to explain allinurl.

4. related- It returns the websites similar to specified websites.

examples :-
related:www.mobivox.com , now mobivox.com provides free/cheap calls facility. This search query would return results containing websites which provide such kind of services.
related:www.hidemyass.com , now hidemyass.com provides services to maintain anonymity using proxies.This search query would return results containing websites which provide services related to proxies.

5. cache: It returns the cached webpage that is kept with google.

example:-
cache:www.explorehacking.com, this dork is useful very when actually the website is down and
you can still view its contents (from cached pages).

 6. ext- It specifies the extension .

You can use 'filteype' at place of 'ext'.
example-:
ext:ppt hacking - this will return  powerpoint presentations (ppt files) on topic 'hacking'.
ext:doc hacking - this will return microsoft word files on topic 'hacking'

 Extensions that google supports are pdf,swf,rtf,doc,ppt,dwf,ps,kml,kmz,xls.


Note: Undoubtedly,you can combine these search operators for example
site:gov inurl:adminlogin will return the government site webpages who have word 'adminlogin' in URL.



Accessing Unprotected Cameras using google ?
So the heading sounds interesting. We can access the live cameras using google dorks.
Remember, cameras are also present at beaches ;)
Just type this in google search box and hit enter.
inurl:view/index.shtml .

Many of these wont ask for password, view might be dark (coz it would be night there) and you need good internet speed . Have fun.

Download a long list of such dorks to view unprotected cameras and the software from here.

Note:These google dorks when used smartly are really useful from hacker's point of view to search for vulnerable websites. I will try to write a post regarding same.
»»  read more

How does Google Search work ?

Google search engine is undoubtedly most widely used search engine. It was founded by Larry Page and Sergey Brin. We must have the knowledge of basic working and methodology used by google search engine.  I have explained the things in very simple words.  Read Carefully

Overview :

Okay lets assume , you wanna design a little search engine that would search the requested key words in  few websites (say 5 websites) ,So what would be our approach ? First of all, we will store the contents that is webpages of that 5 websites in our database. Then we will make an index including the important part of these web pages like titles,headings,meta tags etc. Then we would make a simple search box meant for users where they could enter the search query or keyword. User's entered query will be processed  to match with the keywords in the index and  the results would be returned accordingly. We will return user with list of the links of actual websites and the preference to those websites will be given to them using some algorithm.   I hope the basic overview of  working of search engine is clear to you. 

Now read more regarding the same.


A web search engine works basically in the following manner. There are basically three parts.

1. Web Crawling 
2. Indexing 
3. Query processing or searching

1. First step of working of search engine is web crawling. A web crawler or a web spider is a software that travels across the world wide web and downloads,saves webpages. A web crawaler is fed with URLs of websites and it starts proceeding. It starts downloading and saving web pages associated with that websites. Wanna have feel of web crawaler. Download one from here. Feed it with links of websites and it    will start downloading  webpages,images etc associated with those websites. Name of google web crawler is GoogleBot.  Wanna see the copies of webpages saved in google database ? (actually not exactly)
Lets take example of any website , say www.wikipedia.org
Do this -:
Go to google. and  search for 'wikipedia' Hopefully you would get this link on top.
Click on the 'cached' link as shown.


OR

Directly search for 'cache:wikipedia.org'



Then read the lines at top the page you got and things would be clear to you.

2. After googlebot has saved all pages, it submits them to google indexer. Indexing means extracting out words from titles,headings,metatags etc.The indexed pages are stored in google index database. The contents of index database is similar to the index at the back of your book. Google ignores the common or insignificant words like as,for,the,is,or,on (called as stop words) which are usually in every webpage. Index is done basically to improve the speed of searching.

3. The third part is query processing or searching. It includes the search box where we enter the search query/keyword for which we are looking for. When user enters the serach query, google matches the entered key words in the pages saved in indexed database and returns the actual links of webpages from where those pages are reterived. The priority is obviously given to best matching results. Google uses a patented algorithm called PageRank that helps rank web pages that match a given search string.


The above three steps are followed not only google search but most of the web search engines.Ofcourse there are many variations but methodology is same

What is Robots.txt ?
Web Administrators do not the web crawlers or Web spiders to fetch every page/file of the website and show the links in search results.Robots.txt is a simple text file meant to be placed in top-level directory of the website which contain the links that web administrators do not want to be fetched by web crawlers. The first step of a Web Crawler is to check the content of Robots.txt

Example of contents of Robots.txt
User-agent: * //for web crawlers of all search engines
Disallow:/directory_name/file_name //specify a file of particular dir.
Disallow:/directory_name/  //all files of particular dir.

You can see robots.txt of  websites (if exists). Example http://www.microsoft.com/robots.txt


»»  read more