How To Avoid Being Crawled/penalized By Google
I'd like to test some Django websites using random data on a production server using real domain names, but these websites will be simple tests with possible duplicated data (quite probable not following Google rules). I know usually for this you use a development/staging/virtual box for such a task, but I do want to use directly the production box with the real DNS.
Now, I'm kind of new on website development and SEO, and I wouldn't like to mess with SEO and Google.
What'd be the right way to proceed here? Should I try to avoid being indexed/crawled by Google somehow? Any other advices?
You can disallow the complete indexing of your page and then later (when you're done with coding) activate it again.
Thats probably the best way because google is not going to crawl "bad pages" (for example when your website still is in development) and thus you will not get a Ranking so far (called Page Ranking if you want to look it up).
- → Django, code inside <script> tag doesn't work in a template
- → Uncaught ReferenceError: Parent is not defined
- → React - Django webpack config with dynamic 'output'
- → Put a Rendered Django Template in Json along with some other items
- → Implement shopify templates in django
- → Python Shopify API output formatted datetime string in django template
- → How to avoid being crawled/penalized by Google
- → Django: Identify the urls that provide duplicate content and set a canonical link
- → Shopify app: adding a new shipping address via webhook
- → Jquery Modal Confirmation on Django form submit for deletion of object
- → changing the size of an image with css
- → shopify_auth multi store session handling
- → How to use Shopify Python API RecurringApplicationCharge