Возраст домена | 22 года |
Дата окончания | Истек срок регистрации |
PR | 3 |
ИКС | 0 |
Страниц в Google | 201 |
Страниц в Яндексе | 1 |
Dmoz | Нет |
Яндекс Каталог | Нет |
Alexa Traffic Rank | 18455575 |
Alexa Country | Нет данных |
История изменения показателей | Авторизация |
Идет сбор информации... Обновить
WWW.Smythies.com - Entry point Web Page. 2010.10.14
n/a
The Smythies.com server is physically located at Doug Smythies house in Coquitlam B.C., Canada. It is largely a picture gallery, but also contains both useful and useless information.
UTF-8
9.5 КБ
751
5 255 симв.
4 258 симв.
Данные предоставлены сервисом semrush
Сайт | Общие фразы | PR | тИЦ | Alexa Rank | Alexa Country | |
---|---|---|---|---|---|---|
repairclinic.com | 11 | 5 |
0 | 19317 | 3948 | |
justanswer.com | 11 | 6 |
0 | 5569 | 1451 | |
appliancepartspros.com | 10 | 3 |
0 | 38957 | 7128 | |
manualslib.com | 7 | 4 |
0 | 3436 | 1847 | |
fixya.com | 6 | 5 |
0 | 3502 | 1379 | |
dacor.com | 6 | 4 |
0 | 392582 | 123855 | |
manualsonline.com | 5 | 4 |
0 | 51710 | 22776 | |
ereplacementparts.com | 5 | 4 |
0 | 26470 | 4419 | |
appliancetimers.com | 5 | 5 |
0 | 1838804 | 432541 | |
ajmadison.com | 5 | 4 |
0 | 39959 | 6578 | |
toic.org | 2 | 2 |
0 | 4110860 | Нет данных | |
thegeekstuff.com | 2 | 5 |
0 | 36309 | 7839 | |
techrepublic.com | 2 | 8 |
0 | 3769 | 1979 | |
stackoverflow.com | 2 | 7 |
0 | 47 | 44 | |
stackexchange.com | 2 | 6 |
0 | 118 | 66 | |
raymii.org | 2 | 3 |
20 | 214688 | 177721 | |
marksanborn.net | 2 | 3 |
10 | 836919 | 357683 | |
howtoforge.com | 2 | 6 |
0 | 13530 | 11255 | |
bbrinck.com | 2 | 2 |
0 | 14006746 | Нет данных | |
alexonlinux.com | 2 | 3 |
10 | 1714486 | Нет данных | |
Еще 30 сайтов после авторизации |
Данные предоставлены сервисом semrush
Данные linkpad ( 9 Ноября 2014 ) | |
Количество ссылок на сайт | 274 |
Количество доменов, которые ссылаются на сайт | 3 |
Количество найденных анкоров | 3 |
Исходящие (внешние) ссылки домена | 14 |
Количество доменов, на которые ссылается сайт | 12 |
Количество исходящих анкоров | 11 |
Внутренние ссылки главной страницы ( 43 ) | |
/bot_trap.html | Do not click on this link. It is a trap for bad crawlers. |
smythies.com/ | Go to Smythies.com (absolute) |
./index.html | home page (relative) |
./about_smythies.html | About Smythies.com |
./disclaimer.html | Disclaimer |
./network/index.html | Network (diagrams, status, ...) |
./~doug/genealogy/index.html | Genealogy web pages |
./carrie/index.html | Carrie's web pages |
./~doug/index.html | Doug's web pages |
./graham/index.html | Graham's web pages |
./graham/black_tusk_2011/index.html | Doug and Graham's Black Tusk hike 2011 |
./xmas_2009/index.html | Christmas 2009 |
./xmas_2008/index.html | Christmas 2008 |
./xmas_2006/index.html | Christmas 2006 |
./xmas_2004/index.html | Create Xmas 2004 page. |
./xmas_2003/index.html | Christmas 2003 |
./xmas_2002/index.html | Christmas 2002 |
./bel_haven_2004/index.htm | Bel-Haven 2004 |
./bel_haven_2003/index.htm | Trip to Bel_Haven August 2003. |
./mom85/index.htm | Mom's 85th birthday (bulk processed). |
./gerry_gordon_50th_web/index.htm | Gordon and Gerri Ormistion 50th anniversary |
./gerri_70/index.htm | Pictures from Gerri's 70th birthday. |
./unused_link.html | VE7NP - 56k baud 440Mhz |
./field_hockey/index.html | Field Hockey (Carrie, Cydney) |
./lacrosse/index.html | Lacrosse (Graham) |
./soccer/index.html | Soccer (Carrie, Graham) |
./digital_camera/index.html | Add link to page test/learn page. |
./lab/index.html | Lab project |
./poweredby.html | Various "powered by" plugs |
./test/index.html | Testing area |
./old_pages.html | Old versions of this page |
./index_0049.html | Previous version of this page |
./lacrosse/Midget_A1_Adanacs/index.html | Added Midget A1 Lacrosse team page. |
./family/2004.08.04/index.htm | Pictures from Colbourne family picnic, August 2004 (Cyd's side). |
./graham/rugby/2004.04.07/index.htm | Pictures from Graham's rubgy game April 7th. |
./graham/rugby/2004.03.31/index.htm | Pictures from Graham's rubgy game today. |
./lacrosse/Midget_A2_Adanacs/index.html | Added Midget A2 Lacrosse team page. |
./carrie/kyak_6_web/index.htm | Automatically generated version of Como Lake grade 6, group 2, Ocean Kyak pictures. |
./carrie/kyak_6/index.html | Como Lake grade 6, group 2, Ocean Kyak pictures. |
./field_hockey/crushers/index.html | Added link to U12 Crushers, Field Hockey. |
./lacrosse/Bantam_A1_Adanacs/index.html | Added link to Lacross Boys Bantam A1 pages. |
./dana_broken_nose/index.html | Click here for the pictures of poor Dana with her broken nose. |
./test/table.html | Test versions. |
Domain Name: SMYTHIES.COM
Registry Domain ID: 81900054_DOMAIN_COM-VRSN
Registrar WHOIS Server: whois.domaindiscover.com
Registrar URL: http://www.domaindiscover.com
Updated Date: 2015-11-03T15:46:59Z
Creation Date: 2002-01-03T05:52:27Z
Registry Expiry Date: 2020-01-03T05:52:27Z
Registrar: TierraNet Inc. d/b/a DomainDiscover
Registrar IANA ID: 86
Registrar Abuse Contact Email: abuse@tierra.net
Registrar Abuse Contact Phone: 858-560-9416
Domain Status: clientTransferProhibited https://icann.org/epp#clientTransferProhibited
Name Server: NS1.MEGANAMESERVERS.COM
Name Server: NS2.MEGANAMESERVERS.COM
URL of the ICANN Whois Inaccuracy Complaint Form: https://www.icann.org/wicf/
>>> Last update of whois database: 2018-10-25T09:10:50Z <<<
For more information on Whois status codes, please visit https://icann.org/epp
NOTICE: The expiration date displayed in this record is the date the
registrar's sponsorship of the domain name registration in the registry is
currently set to expire. This date does not necessarily reflect the expiration
date of the domain name registrant's agreement with the sponsoring
registrar. Users may consult the sponsoring registrar's Whois database to
view the registrar's reported date of expiration for this registration.
TERMS OF USE: You are not authorized to access or query our Whois
database through the use of electronic processes that are high-volume and
automated except as reasonably necessary to register domain names or
modify existing registrations; the Data in VeriSign Global Registry
Services' ("VeriSign") Whois database is provided by VeriSign for
information purposes only, and to assist persons in obtaining information
about or related to a domain name registration record. VeriSign does not
guarantee its accuracy. By submitting a Whois query, you agree to abide
by the following terms of use: You agree that you may use this Data only
for lawful purposes and that under no circumstances will you use this Data
to: (1) allow, enable, or otherwise support the transmission of mass
unsolicited, commercial advertising or solicitations via e-mail, telephone,
or facsimile; or (2) enable high volume, automated, electronic processes
that apply to VeriSign (or its computer systems). The compilation,
repackaging, dissemination or other use of this Data is expressly
prohibited without the prior written consent of VeriSign. You agree not to
use electronic processes that are automated and high-volume to access or
query the Whois database except as reasonably necessary to register
domain names or modify existing registrations. VeriSign reserves the right
to restrict your access to the Whois database in its sole discretion to ensure
operational stability. VeriSign may restrict or terminate your access to the
Whois database for failure to abide by these terms of use. VeriSign
reserves the right to modify these terms at any time.
The Registry database contains ONLY .COM, .NET, .EDU domains and
Registrars.
# robots.txt for www.smythies.com
#
# A note to wordpress users that fetch this file as
# an example via the "Multipart robots.txt editor"
# plugin.
# I deny a lot of bots, you might want to re-consider
# for your application.
# Also, this file is only one step in my access management.
# I also use:
# Apache rewrite rules, i.e. for bots that ignore this file.
# Direct iptables DROP rules for annoying crawlers without an
# otherwise uniquely identifiable user agent string.
#
#
# robots.txt 2018.10.30
# dissallow: User-agent: serpstatbot
# dissallow: User-agent: Datanyze
# dissallow: User-agent: IndeedBot
# dissallow: User-agent: Experibot
# dissallow: User-agent: Seekport
# not sure if Datanyze checks robots.txt
#
# robots.txt 2018.08.06
# add .svg files to disallow lists.
#
# robots.txt 2018.05.09
# disallow: User-agent: dataprovider
# disallow: User-agent: crawler4j
# disallow: User-agent: ExtLinksBot
# disallow: User-agent: The Knowledge AI
# Add more version of Sogou crawlers
#
# robots.txt 2018.04.11
# I often post .csv files and .data files for my
# experiments. Add to disallow list
#
# robots.txt 2018.04.10
# disallow: User-agent: MauiBot
# disallow: User-agent: DAUM
#
# robots.txt 2017.12.02
# Change to multiple User-agent: lines per disallow, as
# specified in the original robots.txt specifications.
# disallow: User-agent: SEOkicks-Robot
#
# robots.txt 2017.07.01
# disallow: User-agent: SiteExplorer Findxbot GarlikCrawler ZoominfoBot BUbiNG
# Barkrowler rogerbot dotbot JamesBOT Contacts-Crawler CCBot IDBot DnyzBot
# PiplBot AlphaBot AlphaSeoBot AlphaSeoBot-SA
#
# robots.txt 2017.04.20
# disallow: User-agent: Qwantify
#
# robots.txt 2017.04.15
# 007ac9.net crawlers are seriously annoying.
# While they do attempt to fetch the robots.txt file,
# they do not use any user agent string when they do so,
# and therfore they hit my "no user agent" rewrite rule.
# They use a user agent string for everything else, but
# do not identify themselves as a 007ac9 crawler, making
# a user agent based rewrite rule impossible.
#
# disallow: User-agent: coccocbot-web
#
# robots.txt 2017.03.18
# disallow: User-agent: 007ac9
# although their bots do not identify themselves,
# they claim to honor robots.txt. We'll see.
#
# robots.txt 2017.02.17
# disallow: User-agent: yoozBot
# googlebot does ignore crawl delay. Take it out.
#
# robots.txt 2017.01.20
# googlebot crawls too fast.
# Try a crawl delay directive for googlebot,
# although I seem to recall it ignores it.
#
# robots.txt 2017.01.09
# disallow: User-agent: DomainCrawler
#
# robots.txt 2016.12.28
# disallow: User-agent: Cliqzbot
#
# robots.txt 2016.12.08
# disallow: User-agent: Seeker
#
# robots.txt 2016.09.04
# disallow: User-agent: Uptimebot
#
# robots.txt 2016.08.04
# disallow: User-agent: Sogou web spider
#
# robots.txt 2016.07.08
# I have been watching Qwantify.
# It gets the same stuff over and over again,
# rather often. Try a crawl delay. If that
# doesn't help, I'll just disallow it.
#
# robots.txt 2016.04.03
# disallow: User-agent: RankActiveLinkBot
#
# robots.txt 2016.02.23
# Aboundex has already been added, it turns out twice.
# try Abountdex only, in addition to the already
# rule for Abountdexbot.
# Will also be adding a re-write rule.
#
# robots.txt 2016.02.09
# disallow: User-agent: plukkie
# disallow: User-agent: Applebot
# disallow: User-agent: Lipperhey
# disallow: User-agent: SafeDNSBot
#
# robots.txt 2016.01.09
# Try, does this work?
# disallow: User-agent: gocrawl
#
# robots.txt 2015.10.25
# disallow: User-agent: NextGenSearchBot
#
# robots.txt 2015.10.17
# disallow: User-agent: parsijoo-bot
#
# robots.txt 2015.10.09
# disallow: User-agent: betaBot
#
# robots.txt 2015.09.13
# disallow: User-agent: RankSonicBot
#
# robots.txt 2015.09.08
# try this instead:
# disallow: User-agent: yacybot
#
# robots.txt 2015.09.08
# disallow: User-agent: YaCy
# I do not know if it should be
# YaCy or yacybot.
#
# robots.txt 2015.08.25
# disallow: User-agent: thunderstone
#
# robots.txt 2015.08.19
# The Nutch disallow is not working.
# The apache web site says it should work.
# Try "tbot-nutch".
# Oh my god, these bots are annoying.
#
# robots.txt 2015.07.20
# Google Search Appliance
# disallow: User-agent: gsa-crawler
# Apache Nutch-based bots.
# dissallow: User-agent: Nutch
#
# robots.txt 2015.07.10
# disallow: User-agent: LSSRocketCrawler
#
# robots.txt 2015.06.20
# disallow: User-agent: YisouSpider
#
# robots.txt 2015.05.29
# Yet another demented bot.
# disallow: User-agent: SMTBot
#
# robots.txt 2015.01.25
# Add some directives for slurp (Yahoo)
# See also 2009.09.09
#
# robots.txt 2015.01.22
# disallow: User-agent: ltx71
# disallow: User-agent: AdvBot
#
# robots.txt 2015.01.10
# Make sure all know to avoid bot_trap.html
#
# robots.txt 2014.12.31
# disallow: User-agent: memoryBot
# Another day, another challenged bot.
#
# robots.txt 2014.11.22
# disallow: User-agent: MojeekBot
# It doesn't need to check this file for every access.
#
# robots.txt 2014.11.13
# dissallow .mp3
#
# robots.txt 2014.10.08
# There seems to be a lot more Bots lately.
# disallow: User-agent: LoadTimeBot
# disallow: User-agent: oBot
#
# robots.txt 2014.10.02
# disallow: User-agent: Riddler
# disallow: User-agent: A6-Indexer
# Seems to check robots.txt often.
# disallow: User-agent: SemrushBot
# Although, it doesn't seem to check this file anyhow.
#
# robots.txt 2014.09.29
# bingbot has had mental breakdown
# and I'm fed up with it.
# disallow: User-agent: bingbot
#
# robots.txt 2014.09.03
# disallow: User-agent: XoviBot
# Although, it doesn't seem to check this file anyhow.
#
# robots.txt 2014.08.28
# bingbot needs to slow down. Use crawl_delay
#
# robots.txt 2014.05.15
# disallow: User-agent: Aboundexbot
# No: already done. See 2013.09.02
#
# robots.txt 2014.05.15
# disallow: User-agent: BLEXBot
#
# robots.txt 2014.04.18
# disallow: User-agent: wotbox
#
# robots.txt 2015.01.22
# disallow: User-agent: lx71EasouSpider
# disallow: User-agent: SeznamBot
#
# robots.txt 2014.02.16
# disallow: User-agent: LinkpadBot
#
# robots.txt 2014.02.15
# disallow: User-agent: archive.org_bot
# (The WayBack machine).
#
# robots.txt 2013.11.18
# disallow: User-agent: spbot
#
# robots.txt 2013.10.07
# disallow: User-agent: Mail.Ru
# disallow: User-agent: meanpathbot
#
# robots.txt 2013.09.02
# Another day, another mentally challenged crawler.
# disallow: User-agent: Aboundexbot
#
# robots.txt 2013.05.15
# Another day, another mentally challenged crawler.
# disallow: User-agent: netEstate NE Crawler
#
# robots.txt 2013.04.25
# disallow ip-web-crawler.com. It crawls way too fast and while
# it claims to obey robtos.txt directives, it does not.
# If it doesn't obey the disallow, then an iptables drop
# 50.31.96.6 - 50.31.96.12 could be used
#
# robots.txt 2013.04.17
# add some dissallow stuff for specific file extensions.
# Somehow I missed it before.
#
# robots.txt 2013.04.04
# disallow Sosospider. Any web crawler that is too stupid to know the
# difference between upper and lower case is not worthy.
#
# robots.txt 2013.02.28
# disallow Exabot. I wonder if the resulting search engine
# database is the reason I get so many forged referrer
# hits.
#
# robots.txt 2012.10.08
# disallow WBSearchBot.
#
# robots.txt 2012.09.02
# disallow SearchmetricsBot. It is mentally challenged.
#
# robots.txt 2012.05.03
# disallow TurnitinBot. It is mentally challenged.
#
# robots.txt 2012.03.29
# disallow EC2LinkFinder. I do not know if it obeys robots.txt, but I wll try.
# For sure it ignores most robots.txt directives. It copies everything, hogging
# bandwidth.
# It is time to think of a generic deny, to cover all these new bots.
#
# robots.txt 2012.03.13
# disallow SWEBot. It is not polite and disobaeys robots.txt file.
#
# robots.txt 2012.01.29
# disallow aiHitBot
# Try a useragent "InfoPath" and "InfoPath.2" dissallow. (Another MS thing.)
# I am trying to get rid of what appears to be a tracking site.
# 80.40.134.103, .104, .120, seem to track 92.9.131.199 and 92.9.150.29 and ...
# 80.40.134.XXX does read the robots.txt file.
#
# robots.txt 2012.01.04
# SISTRIX crawler does not behave well. It ignores meta tags and some robots.txt directives.
# Disallow it.
#
# robots.txt 2011.12.01
# Try to get rid of Ezooms bot, although it is not clear what its exact user agent name is.
# (Days later: "User-agent: Ezooms" seems to work, but it takes a few days.)
# It ignores meta tags, and has become generally annoying.
#
# robots.txt 2011.09.26
# Until now I have allowed Baiduspider. But it has gone mental and also ignores some meta tags.
# Disallow it.
# A new robot, AhrefsBot, does not behave or obey meta tags.
# Disallow it.
#
# robots.txt 2011.06.19
#
# robots.txt 2011.04.12
# Googlebot is so very very severely mentally challenged.
# It ignores the NOFOLLOW meta tag.
# Try to block useless content from being indexed via, yet another,
# block command.
#
# It is still looking for pages that haven't been there for over a year now.
# (see 2010.04.29)
#
# robots.txt 2010.10.14
# Eliminate crawl delay for Yahoo slurp (see 2007.03.13)
#
# robots.txt 2010.09.20
# TwengaBot is severely mentally challenged. Try global disallow for it.
# Googlebot is still annoying and accessing pages it shouldn't.
#
# robots.txt 2010.04.29
# Googlebot is very severely mentally challenged.
# Add disallow directives for directories that are not even there,
# and haven't been for over 5 weeks now.
# This is merely to try to get around having my request to delete the
# non-existant directories from the search database being denied.
#
# robots.txt 2010.04.16
# Add specific directives for exabot, including a crawl delay.
# Reduce the slurp (Yahoo) crawl delay (which it doesn't seem to obey anyhow).
# Disallow googlebot-image.
#
# robots.txt 2010.04.13
# disallow taptubot, the mobile device crawler
#
# robots.txt 2010.04.01
# Yet another attempt to get web crawlers not to index old versions of index.html files.
# All old version are called index_0???.html.
#
# robots.txt 2010.03.19
# Archives have been moved to a seperate directory. Add disallow directive.
#
# robots.txt 2010.02.10
# The Yandex web crawler behaves in a very strange manor. Block it.
# Ask Robots not to copy PDF files.
#
# robots.txt 2009.12.07
# Fix some syntax based on feedback from http://tool.motoricerca.info/robots-checker.phtml
#
# robots.txt 2009.12.04
# There are still issues with googlebot. I don't want old versions of index.html
# type pages indexed, but I do want the photoshop elements generated pages indexed.
# Try some new directives.
#
# robots.txt 2009.09.09
# Googlebot is not ignoring the rebuilt directory and is obtaining .MOV videos.
# Add some more googlebot specific directives.
#
# robots.txt 2009.07.27
# Googlebot directives are case sensitive. Add .JPG to .jpg ignore directives.
# Googlebot is not ignoring old index pages as global directive indicates to. Try a googlebot
# specific directive.
#
# robots.txt 2009.04.12
# Some robots, for example googlebot, obey global directives as well as googlebot specific directives.
# Other robots, for example slurp (Yahoo) and msnbot, only obey their specific directives.
# The robots.txt standard is rather weak, incomplete, and generally annoying.
# Add tons of the same specific directives to each robot area.
# Try to change no index Christmas pages to include a wildcard.
#
# robots.txt 2008.12.03
# Block the Cuil (twiceler) robot entirely.
#
# robots.txt 2008.11.23
# The majestic robot comes in bursts at a high rate. Just block it.
# The Cuil robot comes to much. Try to slow it down.
#
# robots.txt 2008.07.03
# Now msnbot has started to grab images. Try to stop it.
# Googlebot is grabbing PNG files. Try to stop it.
#
# robots.txt 2007.11.20
# Try to disallow the panscient.com web crawler.
#
# robots.txt 2007.08.23
# Still search engine pages do not agree with contents of robots.txt file.
# Add specific disallow for ~doug/rebuilt.
# - put global user agent lines after specific ones.
# - next will be to repeat global lines in each specific agent area.
#
# robots.txt 2007.05.03
# Now Googlebot has started to grab images. Try to stop it.
# For whatever reason, google is mainly showing my re-built directory. It
# never seems to go back to the higher level page that now has meta tags
# telling it not to index those pages. Put in a global disallow.
# Add some other global disallows, that I got behind on.
#
# robots.txt 2007.03.13
# stupid yahoo slurp comes all the time now. It supports a non-standard delay command.
# so add the command. The web site doesn't state the units of measure.
#
# robots.txt 2007.02.11
# yahoo, slurp seems to now obey the non-standard ignore this type of file wildcard usage
# try it.
#
# robots.txt 2006.12.29
# Delete instructions for directories that don't exist anymore
#
# robots.txt 2004:12:21
# Try to eliminate yahoo.com grabbing images.
# Can only think of global deny.
# Can not find Yahoo name, try one shown below.
#
# robots.txt 2004:11:16
# Try to eliminate alexa.com grabbing images.
# InkTomi comes too often, can them entirely.
#
# robots.txt 2004:07:16
# Try to eliminate picsearch.com grabbing images.
#
# robots.txt 2004:07:09
# Try to eliminate altavista grabbing images.
#
User-agent: serpstatbot
User-agent: Seekport
User-agent: Datanyze
User-agent: Experibot
User-agent: IndeedBot
User-agent: ExtLinksBot
User-agent: crawler4j
User-agent: dataprovider
User-agent: DAUM
User-agent: MauiBot
User-agent: panscient.com
User-agent: vscooter
User-agent: psbot
User-agent: ia_archiver
User-agent: MJ12bot
User-agent: twiceler
User-agent: Yandex
User-agent: taptubot
User-agent: Googlebot-Image
User-agent: TwengaBot
User-agent: sitebot
User-agent: Baiduspider
User-agent: AhrefsBot
User-agent: Ezooms
User-agent: sistrix
User-agent: aiHitBot
User-agent: InfoPath
User-agent: InfoPath.2
User-agent: swebot
User-agent: EC2LinkFinder
User-agent: TurnitinBot
User-agent: The Knowledge AI
User-agent: Mappy
Disallow: /
User-agent: SearchmetericsBot
User-agent: WBSearchBot
User-agent: Exabot
User-agent: Sosospider
User-agent: ip-web-crawler.com
User-agent: netEstate NE Crawler
User-agent: Aboundexbot
User-agent: Aboundex
User-agent: meanpathbot
User-agent: Mail.Ru
User-agent: spbot
User-agent: archive.org_bot
User-agent: LinkpadBot
User-agent: EasouSpider
User-agent: SeznamBot
User-agent: wotbox
User-agent: BLEXBot
User-agent: XoviBot
User-agent: SemrushBot
User-agent: A6-Indexer
User-agent: Riddler
User-agent: LoadTimeBot
User-agent: oBot
User-agent: MojeekBot
User-agent: memoryBot
User-agent: ltx71
Disallow: /
User-agent: AdvBot
User-agent: SMTBot
User-agent: YisouSpider
User-agent: LSSRocketCrawler
User-agent: gsa-crawler
User-agent: Nutch
User-agent: tbot-nutch
User-agent: thunderstone
User-agent: yacybot
User-agent: RankSonicBot
User-agent: betaBot
User-agent: parsijoo-bot
User-agent: NextGenSearchBot
User-agent: gocrawl
User-agent: plukkie
User-agent: Applebot
User-agent: Lipperhey
User-agent: SafeDNSBot
User-agent: RankActiveLinkBot
User-agent: Sogou blog
User-agent: Sogou inst spider
User-agent: Sogou News Spider
User-agent: Sogou Orion spider
User-agent: Sogou spider2
User-agent: Sogou web spider
User-agent: Uptimebot
User-agent: Seeker
User-agent: Cliqzbot
User-agent: DomainCrawler
User-agent: yoozBot
User-agent: 007ac9
User-agent: coccocbot-web
User-agent: Qwantify
User-agent: SiteExplorer
User-agent: Findxbot
User-agent: GarlikCrawler
User-agent: ZoominfoBot
User-agent: BUbiNG
User-agent: Barkrowler
User-agent: rogerbot
User-agent: dotbot
User-agent: JamesBOT
User-agent: Contacts-Crawler
User-agent: CCBot
User-agent: IDBot
User-agent: DnyzBot
User-agent: PiplBot
User-agent: AlphaBot
User-agent: AlphaSeoBot
User-agent: AlphaSeoBot-SA
User-agent: SEOkicks-Robot
Disallow: /
User-agent: Slurp
Disallow: /*.jpg
Disallow: /*.JPG
Disallow: /*.png
Disallow: /*.PDF
Disallow: /*.pdf
Disallow: /*.mp3
Disallow: /*.MOV
Disallow: /*.mov
Disallow: /*.AVI
Disallow: /*.avi
Disallow: /*.svg
Disallow: /*.csv
Disallow: /*.data
Disallow: /disclaimer.html
Disallow: /security.html
Disallow: /poweredby.html
Disallow: /about_smythies.html
Disallow: /unused_link.html
Disallow: /old_pages.html
Disallow: /index_0*
Disallow: /*index_0*$
Disallow: /digital_camera/
Disallow: /lab/
Disallow: /xmas_*
Disallow: /~doug/archives/
Disallow: /~doug/linux/ubuntu-docs/help.ubuntu.com/
Disallow: /bot_trap.html
User-agent: Googlebot
Disallow: /*.jpg$
Disallow: /*.JPG$
Disallow: /*.png$
Disallow: /*.PDF$
Disallow: /*.pdf$
Disallow: /*.mp3$
Disallow: /*.MOV$
Disallow: /*.mov$
Disallow: /*.AVI$
Disallow: /*.avi$
Disallow: /*.csv$
Disallow: /*.svg$
Disallow: /*.data$
Disallow: /index_0*$
Disallow: /*index_0*$
Disallow: /xmas_*
Disallow: /~doug/archives/
Disallow: /~doug/linux/ubuntu-docs/help.ubuntu.com/
Disallow: /~doug/2010.01.23/
Disallow: /~doug/2007.11.20/
Disallow: /~doug/2004.06.26/
Disallow: /digital_camera/
Disallow: /old_pages.html
Disallow: /unused_link.html
Disallow: /disclaimer.html
Disallow: /security.html
Disallow: /about_smythies.html
Disallow: /poweredby.html
Disallow: /DSCN*.htm
Disallow: /bot_trap.html
User-agent: msnbot
Disallow: /*.jpg$
Disallow: /*.JPG$
Disallow: /*.png$
Disallow: /*.PDF$
Disallow: /*.pdf$
Disallow: /*.mp3$
Disallow: /*.MOV$
Disallow: /*.mov$
Disallow: /*.AVI$
Disallow: /*.avi$
Disallow: /*.csv$
Disallow: /*.svg$
Disallow: /*.data$
Disallow: /disclaimer.html
Disallow: /security.html
Disallow: /poweredby.html
Disallow: /about_smythies.html
Disallow: /unused_link.html
Disallow: /old_pages.html
Disallow: /index_0*
Disallow: /*index_0*$
Disallow: /digital_camera/
Disallow: /lab/
Disallow: /xmas_*
Disallow: /~doug/archives/
Disallow: /~doug/linux/ubuntu-docs/help.ubuntu.com/
Disallow: /bot_trap.html
User-agent: bingbot
Disallow: /
# Crawl-delay: 1800
# Disallow: /*.jpg$
# Disallow: /*.JPG$
# Disallow: /*.png$
# Disallow: /*.PDF$
# Disallow: /*.pdf$
# Disallow: /*.mp3$
# Disallow: /*.MOV$
# Disallow: /*.mov$
# Disallow: /*.AVI$
# Disallow: /*.avi$
# Disallow: /*.svg$
# Disallow: /*.csv$
# Disallow: /*.data$
# Disallow: /disclaimer.html
# Disallow: /security.html
# Disallow: /poweredby.html
# Disallow: /about_smythies.html
# Disallow: /unused_link.html
# Disallow: /old_pages.html
# Disallow: /index_0*
# Disallow: /*index_0*$
# Disallow: /digital_camera/
# Disallow: /lab/
# Disallow: /xmas_*
# Disallow: /~doug/archives/
# Disallow: /~doug/linux/ubuntu-docs/help.ubuntu.com/
# Disallow: /bot_trap.html
User-agent: *
Disallow: /*.jpg
Disallow: /*.JPG
Disallow: /*.png
Disallow: /*.PDF
Disallow: /*.pdf
Disallow: /*.mp3
Disallow: /*.MOV
Disallow: /*.mov
Disallow: /*.AVI
Disallow: /*.avi
Disallow: /*.svg
Disallow: /*.csv
Disallow: /*.data
Disallow: /disclaimer.html
Disallow: /security.html
Disallow: /poweredby.html
Disallow: /about_smythies.html
Disallow: /unused_link.html
Disallow: /old_pages.html
Disallow: /index_0*
Disallow: /*index_0*$
Disallow: /digital_camera/
Disallow: /lab/
Disallow: /xmas_*
Disallow: /~doug/archives/
Disallow: /~doug/linux/ubuntu-docs/help.ubuntu.com/
Disallow: /bot_trap.html
Британия - Лондон - 109.74.204.236
Linode
Linode
HTTP/1.1 200 OK
Date: Sat, 12 Oct 2019 11:32:15 GMT
Server: Apache/2.4.18 (Ubuntu)
Last-Modified: Sun, 31 Jan 2016 17:26:30 GMT
ETag: "25ff-52aa491a92df7"
Accept-Ranges: bytes
Content-Length: 9727
Vary: Accept-Encoding
Connection: close
Content-Type: text/html
Кнопка для анализа сайта в один клик, для установки перетащите ссылку на "Панель закладок"