ИКС | n/a |
Страниц в Google | 3440 |
Страниц в Яндексе | 200 |
Dmoz | Нет |
Яндекс Каталог | Нет |
Alexa Traffic Rank | 1022310 |
Alexa Country | ![]() |
История изменения показателей | Авторизация |
Идет сбор информации... Обновить
Wiki de jeuxvideo.com
n/a
n/a
UTF-8
36.67 КБ
693
5 789 симв.
4 863 симв.
Данные предоставлены сервисом semrush
Сайт | Общие фразы | PR | тИЦ | Alexa Rank | Alexa Country | |
---|---|---|---|---|---|---|
![]() |
1 | 3 |
0 | 9238750 | ![]() |
|
![]() |
1 | n/a | 0 | 11018397 | ![]() |
|
![]() |
1 | n/a | 0 | 687819 | ![]() | |
![]() |
1 | n/a | 0 | 15102246 | ![]() |
|
![]() |
1 | n/a | 0 | 13136587 | ![]() |
|
![]() |
1 | n/a | 0 | Нет данных | ![]() |
|
![]() |
1 | n/a | 0 | 4590147 | ![]() |
|
![]() |
1 | n/a | 0 | 27249639 | ![]() |
|
![]() ![]() |
1 | n/a | 0 | Нет данных | ![]() |
|
![]() ![]() |
1 | n/a | 0 | Нет данных | ![]() |
|
Еще 40 сайтов после авторизации |
Данные предоставлены сервисом semrush
#
# LISTE CRAWLERS
# http://robotstxt.org/db.html
# http://robots-txt.com/
# https://www.keycdn.com/blog/web-crawlers
# https://fr.wikipedia.org/robots.txt
# https://github.com/jfqd/robots.txt/blob/master/robots.txt
# https://searchenginewatch.com/sew/news/2067357/bye-bye-crawler-blocking-parasites
# https://yandex.com/support/webmaster/robot-workings/check-yandex-robots.html
#
# robots.txt for http://www.wikipedia.org/ and friends
#
# Please note: There are a lot of pages on this site, and there are
# some misbehaved spiders out there that go _way_ too fast. If you're
# irresponsible, your access to the site may be blocked.
#
# Observed spamming large amounts of https://en.wikipedia.org/?curid=NNNNNN
# and ignoring 429 ratelimit responses, claims to respect robots:
# http://mj12bot.com/
User-agent: MJ12bot
Disallow: /
# Wikipedia work bots:
User-agent: IsraBot
Disallow: /
User-agent: Orthogaffe
Disallow: /
# Crawlers that are kind enough to obey, but which we'd rather not have
# unless they're feeding search engines.
User-agent: UbiCrawler
Disallow: /
User-agent: DOC
Disallow: /
User-agent: Zao
Disallow: /
# Some bots are known to be trouble, particularly those designed to copy
# entire sites. Please obey robots.txt.
User-agent: sitecheck.internetseer.com
Disallow: /
User-agent: Zealbot
Disallow: /
User-agent: MSIECrawler
Disallow: /
User-agent: SiteSnagger
Disallow: /
User-agent: WebStripper
Disallow: /
User-agent: WebCopier
Disallow: /
User-agent: Fetch
Disallow: /
User-agent: Offline Explorer
Disallow: /
User-agent: Teleport
Disallow: /
User-agent: TeleportPro
Disallow: /
User-agent: WebZIP
Disallow: /
User-agent: linko
Disallow: /
User-agent: HTTrack
Disallow: /
User-agent: Microsoft.URL.Control
Disallow: /
User-agent: Xenu
Disallow: /
User-agent: larbin
Disallow: /
User-agent: libwww
Disallow: /
User-agent: ZyBORG
Disallow: /
User-agent: Download Ninja
Disallow: /
# Misbehaving: requests much too fast:
User-agent: fast
Disallow: /
#
# Sorry, wget in its recursive mode is a frequent problem.
# Please read the man page and use it properly; there is a
# --wait option you can use to set the delay between hits,
# for instance.
#
User-agent: wget
Disallow: /
#
# The 'grub' distributed client has been *very* poorly behaved.
#
User-agent: grub-client
Disallow: /
#
# Doesn't follow robots.txt anyway, but...
#
User-agent: k2spider
Disallow: /
#
# Hits many times per second, not acceptable
# http://www.nameprotect.com/botinfo.html
User-agent: NPBot
Disallow: /
# A capture bot, downloads gazillions of pages with no public benefit
# http://www.webreaper.net/
User-agent: WebReaper
Disallow: /
#
# Friendly, low-speed bots are welcome viewing article pages, but not
# dynamically-generated pages please.
#
# Inktomi's "Slurp" can read a minimum delay between hits; if your
# bot supports such a thing using the 'Crawl-delay' or another
# instruction, please let us know.
#
# There is a special exception for API mobileview to allow dynamic
# mobile web & app views to load section content.
# These views aren't HTTP-cached but use parser cache aggressively
# and don't expose special: pages etc.
#
# Another exception is for REST API documentation, located at
# /api/rest_v1/?doc.
#
User-agent: *
Disallow: /api/
Disallow: /trap/
Disallow: /Special:
Disallow: /Special%3A
################################# ROBOTS.TXT ###################################
# #
# Alphabetically ordered whitelisting of legitimate web robots, which obey the #
# Robots Exclusion Standard (robots.txt). Each bot is shortly described in a #
# comment above the (list of) user-agent(s). Uncomment or delete bots you do #
# not wish to allow on your website / which do not need to visit your website. #
# Important: Blank lines are not allowed in the final robots.txt file! #
# Updates can be retrieved from: https://github.com/jonasjacek/robots.txt #
# #
# This document is licensed with a CC BY-NC-SA 4.0 license. #
# #
# Last update: 2019-03-07 #
# #
################################################################################
# so.com chinese search engine
User-agent: 360Spider
User-agent: 360Spider-Image
User-agent: 360Spider-Video
# google.com landing page quality check
# User-agent: AdsBot-Google
# google.com app resource fetcher
# User-agent: AdsBot-Google-Mobile-Apps
# bing ads bot
User-agent: adidxbot
# apple.com search engine
User-agent: Applebot
user-agent: AppleNewsBot
# baidu.com chinese search engine
User-agent: Baiduspider
User-agent: Baiduspider-image
User-agent: Baidu
User-agent: Baiduspider-video
User-agent: Baiduspider-news
User-agent: Baiduspider-favo
User-agent: Baiduspider-cpro
User-agent: Baiduspider-ads
# bing.com international search engine
User-agent: bingbot
User-agent: BingPreview
# bublup.com suggestion/search engine
User-agent: BublupBot
# commoncrawl.org open repository of web crawl data
User-agent: CCBot
# cliqz.com german in-product search engine
User-agent: Cliqzbot
# coccoc.com vietnamese search engine
User-agent: coccoc
User-agent: coccocbot-image
User-agent: coccocbot-web
# daum.net korean search engine
User-agent: Daumoa
# dazoo.fr french search engine
User-agent: Dazoobot
# deusu.de german search engine
User-agent: DeuSu
# eurip.com european search engine
User-agent: EuripBot
# exploratodo.com latin search engine
User-agent: Exploratodo
# facebook.com social network
User-agent: Facebot
# feedly.com feed fetcher
User-agent: Feedly
# findx.com european search engine
User-agent: Findxbot
# goo.ne.jp japanese search engine
User-agent: gooblog
# so.com chinese search engine
User-agent: HaoSouSpider
# goo.ne.jp japanese search engine
User-agent: ichiro
# istella.it italian search engine
User-agent: istellabot
# jike.com / chinaso.com chinese search engine
User-agent: JikeSpider
# lycos.com & hotbot.com international search engine
User-agent: Lycos
# mail.ru russian search engine
User-agent: Mail.Ru
# google.com adsense bot
# User-agent: Mediapartners-Google
# mojeek.com search engine
User-agent: MojeekBot
# bing.com international search engine
User-agent: msnbot
User-agent: msnbot-media
# orange.com international search engine
User-agent: OrangeBot
# pinterest.com social networtk
User-agent: Pinterest
# botje.nl dutch search engine
User-agent: Plukkie
# qwant.com french search engine
User-agent: Qwantify
User-agent: Qwant
# rambler.ru russian search engine
User-agent: Rambler
# seznam.cz czech search engine
User-agent: SeznamBot
# soso.com chinese search engine
User-agent: Sosospider
# yahoo.com international search engine
User-agent: Slurp
# sogou.com chinese search engine
User-agent: Sogou blog
User-agent: Sogou inst spider
User-agent: Sogou News Spider
User-agent: Sogou Orion spider
User-agent: Sogou spider2
User-agent: Sogou web spider
User-agent: Sogou web
User-agent: Sogou
User-agent: SoGou
User-agent: Sogou Pic Spider/3.0( http://www.sogou.com/docs/help/webmasters.htm#07)
User-agent: Sogou head spider/3.0( http://www.sogou.com/docs/help/webmasters.htm#07)
User-agent: Sogou web spider/4.0(+http://www.sogou.com/docs/help/webmasters.htm#07)
User-agent: Sogou Orion spider/3.0( http://www.sogou.com/docs/help/webmasters.htm#07)
User-agent: Sogou-Test-Spider/4.0 (compatible; MSIE 5.5; Windows 98)
# sputnik.ru russian search engine
User-agent: SputnikBot
# ask.com international search engine
User-agent: Teoma
# twitter.com bot
User-agent: Twitterbot
# wotbox.com international search engine
User-agent: wotbox
# yacy.net p2p search software
User-agent: yacybot
# yandex.com russian search engine
User-agent: Yandex
User-agent: YandexBot
User-agent: YandexMobileBot
# search.naver.com south korean search engine
user-agent: Yeti
# yioop.com international search engine
User-agent: YioopBot
# yooz.ir iranian search engine
User-agent: yoozBot
# youdao.com chinese search engine
User-agent: YoudaoBot
# crawling rule(s) for above bots
Disallow: /
# AJOUTS #
User-agent: Goo
User-agent: Ezooms
User-agent: dotbot
User-agent: AhrefsBot
User-agent: SemrushBot
User-agent: SemrushBot-SA
User-agent: serpstatbot
Disallow: /
User-agent: Exabot
Disallow: /
User-agent: facebot
User-agent: facebookexternalhit
Disallow: /
США - Ашберн - 66.6.41.21
США - Ашберн - 66.6.42.21
США - Ашберн - 66.6.43.21
Tumblr
Tumblr
HTTP/1.1 200 OK
Server: nginx
Date: Wed, 18 Dec 2019 09:20:12 GMT
Content-Type: text/html; charset=UTF-8
Transfer-Encoding: chunked
Connection: keep-alive
Vary: Accept-Encoding
X-Content-Type-Options: nosniff
Content-language: fr
Vary: Accept-Encoding, Cookie
Expires: Thu, 01 Jan 1970 00:00:00 GMT
Cache-Control: private, must-revalidate, max-age=0
Strict-Transport-Security: max-age=0; includeSubDomains
X-Frame-Options: SAMEORIGIN
Access-Control-Allow-Origin: *
X-XSS-Protection: 1; mode=block
Кнопка для анализа сайта в один клик, для установки перетащите ссылку на "Панель закладок"