Published on 2025-08-07T06:18:08Z
TurnitinBot
TurnitinBot is the official web crawler for Turnitin, the well-known academic plagiarism detection service. Its purpose is to scan and collect public, text-based web content to build the vast comparison database that is used to check student submissions for plagiarism. The bot's activity supports academic integrity but does not offer a direct benefit, such as referral traffic, to the websites it crawls.
What is TurnitinBot?
TurnitinBot is the web crawler for Turnitin, a company that provides academic integrity solutions to educational institutions. The bot systematically crawls the internet to collect publicly available content for its plagiarism detection service. It identifies itself in server logs with a user-agent string like TurnitinBot/2.1 (http://www.turnitin.com/robot/crawlerinfo.html)
. The bot is designed to focus specifically on extracting text content and does not process dynamic elements like JavaScript. The content it collects becomes part of Turnitin's reference database.
Why is TurnitinBot crawling my site?
TurnitinBot is visiting your website to gather its text content to add to its comparison database. The bot prioritizes content with academic relevance, such as scholarly articles and educational resources, which students might reference or copy. The frequency of visits depends on your site's content and how often it is updated. The crawling is generally considered authorized as it respects standard web protocols and accesses only publicly available information.
What is the purpose of TurnitinBot?
The purpose of TurnitinBot is to support Turnitin's plagiarism detection service. When a student submits a paper, it is compared against the bot's database to identify potential matches with existing online sources. This helps educational institutions maintain academic integrity. The bot does not store complete copies of websites but creates text fingerprints for similarity matching. While website owners do not directly benefit from the crawl, the broader academic ecosystem benefits from tools that uphold standards of original scholarship.
How do I block TurnitinBot?
To prevent TurnitinBot from indexing your content for its plagiarism database, you can add a specific disallow rule to your robots.txt
file. This is the standard method for managing crawler access.
To block this bot, add the following lines to your robots.txt
file:
User-agent: TurnitinBot
Disallow: /
How to verify the authenticity of the user-agent operated by Turnitin?
Reverse IP lookup technique
host
linux command two times with the IP address of the requester.-
This command returns the reverse lookup hostname (e.g., 4.4.8.8.in-addr.arpa.).> host IPAddressOfRequest
-
> host ReverseDNSFromTheOutputOfFirstRequest