Google Hacking Cheat Sheet Advanced Querying

Total Page:16

File Type:pdf, Size:1020Kb

Google Hacking Cheat Sheet Advanced Querying Google Hacking Cheat Sheet Advanced Querying Query Description & Example inurl: Value is contained somewhere in the url. Ex: “preventing ransomware inurl:fbi” ​ ​ site: Only search within this website’s given domain Ex: “windows xp end of life site:microsoft.com” ​ ​ filetype: Search only for files, not webpages Ex: “nasa moon landing filetype:JPG” ​ ​ allinurl: Search all of the following words in the url. Ex: allinurl: ”blog wordpress” information security ​ ​ intext: Search the body of the webpage for specific text. Ex: patient record intext:"index of /htdocs" ​ related: Find website results that are related to your search term. Ex: related:sans.org ​ info: Find supplemental information Google may have on this page (useful for finding cached pages) Ex: info:www.usgs.gov ​ link: Find other pages indexed by Google that reference this link Ex: link:http://www.somecompany.com/supersecretfile.doc ​ “quote” Find an exact phrase (though results may include related words) Ex: “Malware Hunting” ​ +word Show results with this word exactly. Do not include similar words. Ex: Malware +Hunter ​ ­word/query Do not include this word in search results or queries. Ex: Advanced Malware Hunting ­beginner ­introduction ­site:microsoft.com ​ “word * word” Wildcard. Search for anything between these two words, but include both. Ex: “Next * Firewalls with *” ​ OR (or | ) Return results for either item. The pipe character can be used in place Ex: “locky OR ransomware”; Ex: “locky | ransomware” ​ ​ ​ AND (or &) Return results with both items. Ampersand character can be used in place. Ex: “cissp AND certification” Ex: “cissp & certification” ​ ​ ​ Remember, the real power comes in stringing multiple advanced operators together. (“Index Of” | “[To Parent Directory]”) AND “*financ*” inbody:xlsx site:somebank.com .
Recommended publications
  • Large-Scale, Automatic XSS Detection Using Google Dorks
    Large-Scale, Automatic XSS Detection using Google Dorks Riccardo Pelizzi Tung Tran Alireza Saberi Abstract XSS Attacks continue to be prevalent today, not only because XSS sanitization is a hard problem in rich- formatting contexts, but also because there are so many potential avenues and so many uneducated developers who forget to sanitize reflected content altogether. In this paper, we present Gd0rk, a tool which em- ploys Google’s advanced search capabilities to scan for websites vulnerable to XSS. It automatically generates Figure 1: CVE vulnerabilities for 2010 and maintains a database of parameters to search, and uses heuristics to prioritize scanning hosts which are degrees of client-side support with a primarily server- more likely to be vulnerable. Gd0rk includes a high- side XSS defense [37, 23, 31]. However, the diffusion of throughput XSS scanner which reverse engineers and ap- such methods remains limited: hybrid methods require proximates XSS filters using a limited number of web re- support from both clients and servers. Since the party quests and generates working exploits using HTML and that is most directly affected by an XSS attack is the user JavaScript context-aware rules. who accesses a vulnerable server, client-side protections The output produced by the tool is not only a remark- are thus desirable (despite their limitation to so-called re- ably vast database of vulnerable websites along with flected XSS) and have been developed [29,4, 17, 11]. working XSS exploits, but also a more compact repre- However, client-side defenses are no definitive solution sentation of the list in the form of google search terms, either: IE8 regular-expression based approach is easy whose effectiveness has been tested during the search.
    [Show full text]
  • Recent Developments in Cybersecurity Melanie J
    American University Business Law Review Volume 2 | Issue 2 Article 1 2013 Fiddling on the Roof: Recent Developments in Cybersecurity Melanie J. Teplinsky Follow this and additional works at: http://digitalcommons.wcl.american.edu/aublr Part of the Law Commons Recommended Citation Teplinsky, Melanie J. "Fiddling on the Roof: Recent Developments in Cybersecurity." American University Business Law Review 2, no. 2 (2013): 225-322. This Article is brought to you for free and open access by the Washington College of Law Journals & Law Reviews at Digital Commons @ American University Washington College of Law. It has been accepted for inclusion in American University Business Law Review by an authorized administrator of Digital Commons @ American University Washington College of Law. For more information, please contact [email protected]. ARTICLES FIDDLING ON THE ROOF: RECENT DEVELOPMENTS IN CYBERSECURITY MELANIE J. TEPLINSKY* TABLE OF CONTENTS Introduction .......................................... ..... 227 I. The Promise and Peril of Cyberspace .............. ........ 227 II. Self-Regulation and the Challenge of Critical Infrastructure ......... 232 III. The Changing Face of Cybersecurity: Technology Trends ............ 233 A. Mobile Technology ......................... 233 B. Cloud Computing ........................... ...... 237 C. Social Networking ................................. 241 IV. The Changing Face of Cybersecurity: Cyberthreat Trends ............ 244 A. Cybercrime ................................. ..... 249 1. Costs of Cybercrime
    [Show full text]
  • Hacking the Master Switch? the Role of Infrastructure in Google's
    Hacking the Master Switch? The Role of Infrastructure in Google’s Network Neutrality Strategy in the 2000s by John Harris Stevenson A thesis submitteD in conformity with the requirements for the Degree of Doctor of Philosophy Faculty of Information University of Toronto © Copyright by John Harris Stevenson 2017 Hacking the Master Switch? The Role of Infrastructure in Google’s Network Neutrality Strategy in the 2000s John Harris Stevenson Doctor of Philosophy Faculty of Information University of Toronto 2017 Abstract During most of the decade of the 2000s, global Internet company Google Inc. was one of the most prominent public champions of the notion of network neutrality, the network design principle conceived by Tim Wu that all Internet traffic should be treated equally by network operators. However, in 2010, following a series of joint policy statements on network neutrality with telecommunications giant Verizon, Google fell nearly silent on the issue, despite Wu arguing that a neutral Internet was vital to Google’s survival. During this period, Google engaged in a massive expansion of its services and technical infrastructure. My research examines the influence of Google’s systems and service offerings on the company’s approach to network neutrality policy making. Drawing on documentary evidence and network analysis data, I identify Google’s global proprietary networks and server locations worldwide, including over 1500 Google edge caching servers located at Internet service providers. ii I argue that the affordances provided by its systems allowed Google to mitigate potential retail and transit ISP gatekeeping. Drawing on the work of Latour and Callon in Actor– network theory, I posit the existence of at least one actor-network formed among Google and ISPs, centred on an interest in the utility of Google’s edge caching servers and the success of the Android operating system.
    [Show full text]
  • Google Dorks: Use Cases and Adaption Study
    Google dorks: Use cases and Adaption study UNIVERSITY OF TURKU Department of Future Technologies Master of Science in Technology Thesis Networked Systems Security October 2020 Reza Abasi Supervisors: Dr. Ali Farooq Dr. Antti Hakkala The originality of this thesis has been checked in accordance with the University of Turku quality assurance system using the Turnitin OriginalityCheck service. i UNIVERSITY OF TURKU Department of Future Technologies Reza Abasi: Google dorks: Use cases and adaption study Master of Science in Technology Thesis, 93 pages. Networked Systems Security October 2020 The information age brought about radical changes in our lives. More and more assets are getting connected to the Internet. On the one hand, the connectivity to this ever-growing network of connected devices and assets (the Internet) precipitates more convenience and access to various resources. However, on the downside, the Internet could be the hotbed for malicious actors like hackers, attackers, and cybercriminals’ communities. Continuous Penetration testing and monitoring of the sites, and forums providing illicit digital products and services is a must-do task nowadays. Advanced searching techniques could be employed for discovering such forums and sites. Google dorks that are utilizing Google’s advanced searching techniques could be applied for such purpose. Google dorks could be used for other areas that we will explain during this thesis in more detail like information gathering, vulnerability detection, etc. The purpose of this thesis is to propose advanced searching techniques that will help cybersecurity professionals in information gathering, reconnaissance, vulnerability detection as well as cyber criminal investigative tasks. Further, a usability study has been conducted to examine the acceptance of these techniques among a group of cybersecurity professionals.
    [Show full text]
  • Google Hacking 101
    Google Hacking 101 Edited by Matt Payne, CISSP 15 June 2005 http://MattPayne.org/talks/gh 1 Outline • Google Bombing • Schneier in Secrets and Lies – Attack at a distance – Emergent behavior – Automation • Google as a mirror • “Interesting Searches” – Software versions – Passwords, credit card numbers, ISOs • CGI Scanning – Vulnerable software • Defense against Google Hacking 2 Google Bombing != Google Hacking • http://en.wikipedia.org/wiki/Google_bomb • A Google bomb or Google wash is an attempt to influence the ranking of a given site in results returned by the Google search engine. Due to the way that Google's Page Rank algorithm works, a website will be ranked higher if the sites that link to that page all use consistent anchor text. 3 So What Determines Page Relevance and Rating? • Exact Phrase: are your keywords found as an exact phrase in any pages? • Adjacency: how close are your keywords to each other? • Weighting: how many times do the keywords appear in the page? • PageRank/Links: How many links point to the page? How many links are actually in the page? Equation: (Exact Phrase Hit)+(AdjacencyFactor)+(Weight) * (PageRank/Links) 4 From: Google 201, Advanced Googology - Patrick Crispen, CSU Simply Put • “Google allows for a great deal of target reconnaissance that results in little or no exposure for the attacker.” – Johnny Long • Using Google as a “mirror” searches find: – Google searches for Credit Card and SS #s – Google searches for passwords – CGI (active content) scanning 5 Anatomy of a Search Server Side Client Side http://computer.howstuffworks.com/search-engine1.htm 6 How Google Finds Pages • Are only connected web pages indexed? • NO! – Opera submits every URL viewed to Google for later indexing….
    [Show full text]
  • EVILSEED: a Guided Approach to Finding Malicious Web Pages
    2012 IEEE Symposium on Security and Privacy EVILSEED: A Guided Approach to Finding Malicious Web Pages Luca Invernizzi Paolo Milani Comparetti UC Santa Barbara Lastline, Inc. and Vienna Univ. of Technology [email protected] [email protected] Stefano Benvenuti Christopher Kruegel University of Genova UC Santa Barbara [email protected] [email protected] Marco Cova Giovanni Vigna Lastline, Inc. and University of Birmingham UC Santa Barbara [email protected] [email protected] Abstract—Malicious web pages that use drive-by download In addition to drive-by-download exploits, cybercriminals attacks or social engineering techniques to install unwanted also use social engineering to trick victims into installing or software on a user’s computer have become the main avenue running untrusted software. As an example, consider a web for the propagation of malicious code. To search for malicious web pages, the first step is typically to use a crawler to collect page that asks users to install a fake video player that is URLs that are live on the Internet. Then, fast prefiltering presumably necessary to show a video (when, in fact, it is a techniques are employed to reduce the amount of pages that malware binary). Another example is fake anti-virus programs. need to be examined by more precise, but slower, analysis These programs are spread by web pages that scare users into tools (such as honeyclients). While effective, these techniques thinking that their machine is infected with malware, enticing require a substantial amount of resources. A key reason is that the crawler encounters many pages on the web that are benign, them to download and execute an actual piece of malware as that is, the “toxicity” of the stream of URLs being analyzed is a remedy to the claimed infection- [4], [5].
    [Show full text]
  • Pulp Google Hacking:The Next Generation Search Engine Hacking
    Pulp Google Hacking The Next Generation Search Engine Hacking Arsenal 3 August 2011 – Black Hat 2011 – Las Vegas, NV Presented by: Francis Brown Rob Ragan Stach & Liu, LLC www.stachliu.com Agenda O V E R V I E W • Introduction/Background • Advanced Attacks • Google/Bing Hacking - Core Tools • NEW Diggity Attack Tools • Advanced Defenses • Google/Bing Hacking Alert RSS Feeds • NEW Diggity Alert Feeds and Updates • NEW Diggity Alert RSS Feed Client Tools • Future Directions 2 Introduction/ Background G E T T I N G U P T O S P E E D 3 Open Source Intelligence S E A R C H I N G P U B L I C S O U R C E S OSINT – is a form of intelligence collection management that involves finding, selecting, and acquiring information from publicly available sources and analyzing it to produce actionable intelligence. 4 Google/Bing Hacking S E A R C H E N G I N E A T T A C K S 5 Google/Bing Hacking S E A R C H E N G I N E A T T A C K S Bing's source leaked! class Bing { public static string Search(string query) { return Google.Search(query); } } 6 Attack Targets G O O G L E H A C K I N G D A T A B A S E • Advisories and Vulnerabilities (215) • Pages containing network or • Error Messages (58) vulnerability data (59) • Files containing juicy info (230) • Sensitive Directories (61) • Files containing passwords (135) • Sensitive Online Shopping Info (9) • Files containing usernames (15) • Various Online Devices (201) • Footholds (21) • Vulnerable Files (57) • Pages containing login portals (232) • Vulnerable Servers (48) • Web Server Detection (72) 7 Google Hacking = Lulz R E A L W O R L D T H R E A T LulzSec and Anonymous believed to use Google Hacking as a primary means of identifying vulnerable targets.
    [Show full text]
  • 0321518667 Sample.Pdf
    Many of the designations used by manufacturers and sellers to distinguish their products are claimed as trademarks. Where those designations appear in this book, and the publisher was Editor in Chief aware of a trademark claim, the designations have been printed with initial capital letters or Karen Gettman in all capitals. Acquisitions Editor The author and publisher have taken care in the preparation of this book, but make no expressed or implied warranty of any kind and assume no responsibility for errors or Jessica Goldstein omissions. No liability is assumed for incidental or consequential damages in connection Development Editor with or arising out of the use of the information or programs contained herein. Sheri Cain The views expressed in this article are those of the author and do not reflect the official Managing Editor policy or position of the U.S. Military Academy, the Department of the Army, the Department of Defense, or the U.S. government. Kristy Hart The publisher offers excellent discounts on this book when ordered in quantity for bulk Project Editor purchases or special sales, which may include electronic versions and/or custom covers and Jovana San Nicolas- content particular to your business, training goals, marketing focus, and branding interests. Shirley For more information, please contact: Copy Editor U.S. Corporate and Government Sales Krista Hansing Editorial (800) 382-3419 Services, Inc. [email protected] Indexer For sales outside the United States, please contact: Erika Millen International Sales [email protected] Proofreader Visit us on the web: www.informit.com/aw Kathy Ruiz Library of Congress Cataloging-in-Publication Data available on request.
    [Show full text]
  • Google Dorks: Analysis, Creation, and New Defenses
    Google Dorks: Analysis, Creation, and new Defenses Flavio Toffalini1, Maurizio Abb`a2, Damiano Carra1, and Davide Balzarotti3 1 University of Verona, Italy [email protected], flavio.toff[email protected] 2 LastLine, UK [email protected] 3 Eurecom, France [email protected] Abstract. With the advent of Web 2.0, many users started to maintain personal web pages to show information about themselves, their busi- nesses, or to run simple e-commerce applications. This transition has been facilitated by a large number of frameworks and applications that can be easily installed and customized. Unfortunately, attackers have taken advantage of the widespread use of these technologies { for exam- ple by crafting special search engines queries to fingerprint an application framework and automatically locate possible targets. This approach, usu- ally called Google Dorking, is at the core of many automated exploitation bots. In this paper we tackle this problem in three steps. We first perform a large-scale study of existing dorks, to understand their typology and the information attackers use to identify their target applications. We then propose a defense technique to render URL-based dorks ineffective. Finally we study the effectiveness of building dorks by using only com- binations of generic words, and we propose a simple but effective way to protect web applications against this type of fingerprinting. 1 Introduction In just few years from its first introduction, the Web rapidly evolved from a client-server system to deliver hypertext documents into a complex platform to run stateful, asynchronous, distributed applications. One of the main character- istics that contributed to the success of the Web is the fact that it was designed to help users to create their own content and maintain their own web pages.
    [Show full text]
  • Adversarial Information Retrieval on the Web
    Adversarial Information Retrieval on the Web Marcin Kaminski Motivation Adversarial Information Retrieval has been playing a big role in nowadays Web oriented environment. Almost everybody stores some sensitive content online. Therefore, it is important to know how Adversarial IR can affect us so we can be prepared and successfully protect our confidential data. Introduction During 14th International World Wide Web Conference in Chiba [1], Japan and SIGIR 2006 Conference in Seattle[2] the following subjects on Adversarial Information Retrieval on the Web were discussed: search engine spam and optimization, crawling the web without detection, link-bombing (a.k.a. Google-bombing), comment spam, referrer spam, blog spam (splogs), malicious tagging, reverse engineering of ranking algorithms, advertisement blocking, and web content filtering Introduction (cont) Using web crawlers to steal confidential information. Web Crawlers “A web crawler (also known as a Web spider or Web robot) is a program or automated script which browses the World Wide Web in a methodical, automated manner. Other less frequently used names for Web crawlers are ants, automatic indexers, bots, and worms” [3]. Web Crawlers (cont) a software agent populates the IP addresses on the fly or starts with the predefined list of URL addresses to visit. As it visits these URLs, it gathers all the links it has identified and adds them to the list of URLs to visit, called the crawl frontier . It will then visit these newly added URLs recursively. However, this flow allows some malicious action. a web crawler may reach a sensitive web content which originally was not intended to be available for public view.
    [Show full text]
  • Internet of Torment: the Governance of Smart Home Technologies Against Technology-Facilitated Violence
    Internet of Torment: The Governance of Smart Home Technologies against Technology-Facilitated Violence by Olivia Faria A thesis submitted to the Faculty of Graduate and Postdoctoral Affairs in partial fulfillment of the requirements for the degree of Master of Arts in Communication with specialization in Data Science Carleton University Ottawa, Ontario September 2020, Olivia Faria Abstract While smart home technologies (SHTs) such as smart speakers are often marketed as solutions to automate daily household activities and chores, such as managing calendars, ordering food, playing music or locking the doors, these same technologies can be used to cause harm. This thesis examines the emergence of smart home technology- facilitated violence (smart home TFV) in Canada to assess 1) how these Internet of Things (IoT) devices can be misused as technologies of torment and 2) how and why we may want to examine this phenomenon as part of a wider sociotechnical system of human and non-human actors. First, smart home technologies are situated in the history of home automation and gendered design, conceptually in the research of technology facilitated violence, within a Canadian policy and legal environment and as part of technical research on smart home technology safety by way of an interdisciplinary literature review. Secondly, a modified Walkthrough Method (Light, Burgess and Duguay, 2018) was applied to study the popular Google Nest Mini smart speaker to identify how this device may be misused, which features enable potential misuse and to better understand how the device is reflective of broader corporate values, visions and operating models. Thirdly, a hybrid theoretical framework combining assemblage theory (Kitchin, 2014), actor-network theory (Latour, 1988; Latour, 2005) and multi-scalar analysis (Edwards, 2003) was developed and applied to frame the study of this complex technological system and to structure the collection of observations and analyze results arranged into micro, meso and macro scales.
    [Show full text]
  • Google Chrome
    DIE FRAGE pixabay, Creative Commons CCO Browsen durch das Zwischennetz Colourbox.de BROWSER DER DATENSTROM im Zwischennetz DIE FRAGE SUCHMASCHINEN 5 DIE GESELLSCHAFT DER SUCHE 6 MAN KÖNNTE EINEN BROWSER ALS DAS TOR ZUR DIGITALEN WELT BEZEICHNEN. DENN DIESER IST DER »SCHLÜSSEL« ZUM INTERNET BZW. ZUM WWW, MIT DEM WIR SOWOHL PRIVATE ALS AUCH BERUFLICHE BROWSER AUFGABEN ERLEDIGEN. Nach der Eingabe einer Domain in die Adresszeile des Browsers wird die entsprechende Webseite vom Browser aufgerufen. Nach dem Erreichen der Webseite, enthält der Browser entsprechende Informationen, BROWSER welche Inhalte er laden soll. Was der Browser tut Nachdem der Ladevorgang abgeschlossen ist, stellt der Browser den Inhalt der Webseite in einer für den Nutzer lesbaren Form dar. DER VERRAT DER BROWSER BROWSER Was der Browser weiß https://www.netzwelt.de/tool s/wie-lautet-meine-ip.html Sind die Logindaten kompromittiert? BROWSER Vertrauen ist gut, aber IT-Sicherheit ist kein statischer Zustand, sondern ein laufender Prozess Kuketz-Blog 1. Sicheres Passwort 2. Regelmäßige Updates 3. Zurückhaltende Datenpreisgabe 4. Privatsphäreneinstellungen setzen BROWSER 5. Abzockerfallen erkennen REGELN 6. Sichere Downloads FÜR EINEN SICHEREN UMGANG MIT 7. Vorsicht im Internet NEUEN MEDIEN 8. Gesunden Menschenverstand nutzen 1. Sicheres Passwort 2. Regelmäßige Updates 3. Zurückhaltende Datenpreisgabe 4. Privatsphäreneinstellungen setzen BROWSER 5. Sicheres Surfen asdfghjklöä-:;MNBVCXY 6. Abzockerfallen erkennen 7. Sichere Downloads 8. Vorsicht im Internet 9. Gesunden Menschenverstand nutzen 1. Sicheres Passwort 2. Regelmäßige Updates 3. Zurückhaltende Datenpreisgabe 4. Privatsphäreneinstellungen setzen BROWSER 5. Abzockerfallen erkennen 6. Sichere Downloads 7. Vorsicht im Internet 8. Gesunden Menschenverstand nutzen 1. Sicheres Passwort 2. Regelmäßige Updates 3. Zurückhaltende Datenpreisgabe 4.
    [Show full text]