It's highly likely if that site owner is blocking the commercial tools you mention, they're smart enough to block most others.
But, if money were no option, you could build your own crawler, crawl the entire internet and discover them that way, the it's huge amounts of data and the storage and processing is going to be cost prohibitive most likely.
Or, you could start with someone else who has already done that, like commoncrawl.org, and extract it from their dataset, but it's huge amounts of data and the storage and processing is going to be cost prohibitive most likely (intentionally repeating myself).
OR, like u/TheRealWeedAtman said, invest the above time and effort into your own website, don't worry about the shady tactics of others, and you'll do better in the long term.
The answer is here: https://www.slideshare.net/ipullrank/persona-driven-keyword-research/51-Format_the_rankings_Axis_with
Yep, you gotta do it by hand.
Yep, it's extremely time-consuming.
But it's currently the only way to do it and it's something that's very difficult to automate without a very big margin of error.
​
Yeah I've been using Lighthouse a lot, it's super slick and definitely been nice to be able to test some SEO impacts in UAT before we go live.
I like the article, but unfortunately we're on an ASP.Net framework so some of it is a little different for us (not my circus, not my monkey there). Most of the rest we've already done (sprites, stylesheets, etc.).
Actually got passed this resource from Tom Anthony that provided a pretty good answer on what G is actually looking at: https://moz.com/blog/google-chrome-usage-data-measure-site-speed.
Yep, I applaud /u/timsoulo and ahrefs for taking on the burden of transparency, which has been a core value (quite literally the first core value) at Moz. It has been refreshing to work with Tim and Dmitry in trying to speak the same language of links and indexes and root linking domains. It reminds me of a great post many years ago on Moz that was produced in large part by Dixon Jones at Majestic on Why counting links is hard. It is a great read and as relevant, if not more relevant given canonicals and hreflang, today as it was then.
Let me just share what works for me
>Looking for advice / resources to learn about analyzing technical SEO before starting this job. I saw there’s an SEMRush certificate, anything else I should do? Should I refresh on code?
Those resources are great. study them but also brush up and get exposed to Google search console and chrome console so you can be more granular with your audits.
>Work flow / project managing advice is appreciated too.
Getting things done by david allen is a good start. https://todoist.com/productivity-methods/getting-things-done I know it sounds trivial but you;ll need to up your clarity 10x when working with devs if you actually want to get things done.
Cheers
Is the site server rendered? Like if you "view source" (ctrl-u on Windows or Linux), is there text content there?
Do you have a text browser installed, like lynx? You can see what the pages look like to a bot by using a terminal command like this:
lynx https://example.com/
hmm... not for me - it's OK
I will duplicate link here https://www.slideshare.net/aleydasolis/what-makes-your-seo-fail-and-how-to-fix-it-brightonseo
There are a couple recent ones:
The one I'm actually working on today for a session I'll give tomorrow about how to develop SEO audits that drive growth should be really good too :D I'll need to translate it to English and re-purpose it into a blog post.
This idea that anchor text is only counted from the first link to a page comes from a test Rand Fishkin did over a decade ago (source). Unless you have a more recent example to cite, it is very possible this isn't true anymore. It's kind of a fringe optimization anyways.
Regardless, nearly every website has a logo linking to the homepage in the header, so it can't be true that it's bad for SEO in any way.
Important in international SEO, ccTLDs are the single strongest way to show search engines and users that site content is specifically targeted to a certain country or region — but, importantly, NOT specifically a certain language. When a site uses a ccTLD, Google assumes that site (and all the content on it) is specifically relevant to the geographic area targeted by the ccTLD and should appear on SERPs in that area.
I work at an agency and I'd basically say, choose a firm that doesn't overpromise, pick an agency with transparency, open communication, and a good track record. Also, pick a place that cares about its team and its clients! How many clients does each strategist get -- some agencies give too much work for one strategist and their work gets watered down.
Look at Moz's recommended agency list (full disclosure, my team is on there) but Moz does a great job of picking out the best. You'll get great results with any of those recommended agencies.
AMP pages are cached in Google's servers. You can't just remove them from your own server and expect never to show up again in SERP. To properly remove AMP pages you need to ping Google's AMP cache server and notify the removal of each individual page.
Moving your site to a self hosted account involves making changes to your DNS. Instructions on how to move your domain to a Registrar of your choice are here. You can also edit the DNS on WordPress.com for things like verifying with Google.
Did you Google "Drupal URL alias"? Because, this is the first result (https://www.drupal.org/node/120631):
> ...Drupal has a feature called "URL Alias" that allows you to provide a more understandable name to the content. As far as browsers, servers, and search engines go, it is totally unnecessary. But for humans, it is nearly mandatory. This is why most consultants tell people to always turn on the Path core module, which supports URL aliasing.
> So, just before you submit that treatise on macadamia history, and if you have the Path module enabled (and you will want it), then you'll see a section on the edit page that says "URL path settings." So let's say you want your visitors to see it as http://www.example.com/MacadamiaHistory. In the URL field, enter MacadamiaHistory...
An alias, by definition, is an alternate name for something. In this case, an altername URL/path for your content which exists on nodes by default. If the same content/page is accessible from two different URLs, then you have duplicate content.
An Alias URL is not necesarily SEO-friendly. An alias URL could just be a string of random characters. What makes a URL "seo-friendly" is the inclusion on relevant/contextual keywords and (to some extent) human readability.
So, if the page is about blue shoes you can give it an alias of .../blue-shoes instead of the standard URL the drupal gives (.../node/12345)
E.A.T. as a ranking factor is a myth and Cyrus Shepard doesn't deny it.
Actually, it doesn't matter if it's a ranking factor or not.
It should be the goal of anybody who wants to level up from basic SEO.
Here is how I see it :
Technical, Content and Popularity is the usual SEO Triptych.
By using E.A.T. as the new S.E.O. triptych, it helps to level the ambition:
Before :
Content is all about landing pages. For bonus points, open up a blog.
Technical consists of running an audit to "optimize" for bots.
Popularity is about building links.
After :
Expertise : are you an expert AND can you demonstrate your expertise in an interesting fashion?
Authority : are people influent in your industry talking about you, including your competition? (It's my personal my interpretation of authority, which is actually more about "notoriety")
​
Trust : can we trust your website to be safe, good UX, etc?
There are a lot of great resources out there. This has some links to some solid Technical SEO guides, but I'd also really recommend doing absolutely everything recommended by David Sottimano in this post. You don't need to become a full stack developer to become a technical SEO, but it sure as shit will help.
Always let the user choose the language/location/etc of their choice. Having the 3 language versions of the site is great! (make sure to add hreflang tags now, and using example.com as x-default). I wouldn't redirect users that land on example.com, let them choose what language they want. If they chose, then using the cookie setting to direct them moving forward is ok.
Also, put a language selector in top nav or footer for users to be able to switch languages.
You are definitely on the right path here though. Read this for more info on international seo https://moz.com/learn/seo/international-seo it has other useful links at the bottom.
Devs: Expect them to know nothing if you want to be safe. But generally, developers should grasps the underling concepts of SEO quicker than most. Developers can understand how crawling, indexing and ranking work in a more technical way than a marketing person. So talk to developers, see what they know. Moz has this nice SEO developer cheat sheet that I think developers like.
App Store Optimization: Yes. It is a real thing. It is like SEO from 15 years ago in terms of being much more basic. But yea, people spend a lot of time, resources and money on ASO.
Not to be too much of an a-hole. But you probably won't find too much help here. Most people that participate on this subreddit are SEO professionals. You might find tricks of the trade of you start exploring the subreddit. But giving you a "traffic plan" is what most of us get paid for in one way or another.
That said, SEO is a very self service trade and CAN be learned by webmasters and business owners.
https://moz.com/beginners-guide-to-seo
https://ahrefs.com/blog/seo-basics/
I suggest starting there and then doing some Google searches. If you run across concepts that you're confused with in there, or need help understanding or implementing. Asking a question on this subreddit might begin to bare fruit.
Hope that helps.
Can you do this:
If that doesn't work after a week, then can you setup a sitewide redirect rule (for Googlebot UA only) to 301 to the production URL? Then submit recrawl in GSC?
FYI if devs/sysadmins are being unhelpful - do you have Cloudflare? This is a prime scenario where "edge SEO" would work wonders (if you know what you're doing)
You can Google the IP to see if there are spam reports or if it's a desirable bot.
If you need to verify the IP is who it says it is, you can also do a reverse lookup using a tool like this: https://mxtoolbox.com/ReverseLookup.aspx
In the case of that hit you pasted, it looks like a spambot from Russia. Its IP could be blocked if you want. There is no SEO impact at this stage (as you are no doubt aware).
In the case of the new blog posts, it's impossible to guess why they aren't indexed. If you can PM me the URLs I can take a look.
Always hard to diagnose stuff like this, especially during massive updates. Did you historically redirect 404's to the homepage or anything like that?
Is it a relatively new page? I've seen a lot of people complaining about indexing issues recently.
You might want to check resources for the page as that might be causing Google to just ignore the page or see it as too large to index/cache (look at this file, nearly 7mb! - https://cdn.shopify.com/s/files/1/0049/3443/4888/articles/Cannacares_x_Eva.png?v=1625826586)
On a side not, I'm sure I would use both organisation and website schema on every page. If your going to be placing schema on there why not look at some product specific pieces?
Other platforms already do this, but the mistake I see wordpress making here is assuming tag pages make good pages to have indexed. Glad to see this is only for wordpress.com & not self-hosted .org.
Hi ric_eseeoh,
​
Keyword tools can be really tough. There are a lot of options out there like google ads keyword tool (free), ahrefs, moz. But after years of trying different tools, I really like, and have stuck with: https://semrush.com
Hi Tyler, there's a great tool called Figma. It's very easy to use.
Here's an example of the menu outline I've done - https://www.figma.com/file/iVl2hHGP9MI0Y5AgOQOcLA/Untitled?node-id=0%3A1.
I've been using this tool for years in order to quickly outline my ideas and communicate them to the professional designer who would then create very cool designs. I am a content writer and an emerging SEO, but I am not a designer. Doing something like this in Sketch or PS would surely be a big problem for me, but Figma has been serving me real good!
I could give you some tips in DM with regard to the problem if you need to ;)
I would say SEOwl because it helps me get actionable information AND monitor progress at the same time. What I mean is that I can use it as a reporting/monitoring tools (like many others) but I can also use it to give me insights on the why a SERP change occured (for example : did my competitor changed its meta-description?) using their serp diff tool
If you don't know what you doing switching over to Yoast may cause you issues as the meta data and settings probably aren't sent across. If you just want to enable caching in your WordPress setup you could look at the following plugins https://kinsta.com/blog/wordpress-caching-plugins/ I've personally used WP Super Cache and found it pretty good.
For more info on CDN take a look at this article from Cloudflare which itself is very popular. https://www.cloudflare.com/learning/cdn/what-is-a-cdn/
It stands for cross origin resource sharing and yes, it is something that should be fixed as soon as possible. Usage of CDNs is not all that new and can be beneficial for SEO since response times and user experience are generally better. &nspb; I am surprised the developers didn't offer any solutions since they should be aware that the response headers of the various servers involved are required to forward certain information detailing, among other things, the origin domain of the requested resources. I would initially contact the support services of Cloudflare since this is probably a common occurrence if servers are incorrectly Set up. The following link gives a decent explanation of the problem.
Do you know any JavaScript or react? If so it shouldn’t be to hard to pick up.
Here are their docs
https://nextjs.org/docs/getting-started
If you want a video tutorial series. Maybe check out this guys videos.
I wrote about your problem at https://www.pulno.com/blog/tutorial/seo-multi-language-hreflang. You can look at the source code to see how it is done on Pulno. There's a global version, Polish one, and x-default. Just copy Pulno's solution and it will work perfectly.
You can just use the site or use their chrome extension.
Under the framework and/or Javascript library sections you might see what is being used. For example Forbes.com is built on Angular https://builtwith.com/forbes.com
There might be a daily search limit on builtwith's free version I am not 100% sure, but i have the chrome extension if i am ever interested in the technology a certain site has
Basically any level that requires some kind of user interaction: UGC, private login area, dynamic features, ecommerce, even forms.
In Jekyll there's no place for such features and you are required to add third party plugins for everything (see this page: https://jekyllrb.com/docs/resources/).
So, if it's just a simple addon, I'd see no problems.
But if you need to christmas tree your website with external pugins, then Jekyll shouldn't be your best choice.
I already replied in your newer topic but just to add:
> So, basically there are like 3 hops before the content is read....
This is not how search bots crawl. They look at (internal) links to discover new pages and don't crawl all subdirectories. This is because many sites fake directories.
Look at this random Amazon PDP:
https://www.amazon.com/VTech-Kidizoom-Smartwatch-Amazon-Exclusive/dp/B072JXVTKT/
The folders '/VTech-Kidizoom-Smartwatch-Amazon-Exclusive/and
/VTech-Kidizoom-Smartwatch-Amazon-Exclusive/dp`/ return 404's. They are not directories on categories.
Use canonical tags. Any page with a canonical tag will be version that Google shows in SERPs.
See this random Amazon URL I pulled up.
https://www.amazon.com/Steve-Madden-Norwich-Oxford-Black/dp/B078WMZ2WN/ref=lp_18637582011_1_1?srs=18637582011&ie=UTF8&qid=1550793959&sr=8-1
In the source code, they include this line:
<link rel="canonical" href="https://www.amazon.com/Steve-Madden-Mens-Norwich-Oxford/dp/B078WNZJPJ" />
That is what Google will show.
What you are using right is a proxy service which only works with browsing. You need a VPN service which uses TAP Adapter which overrides the computer IP address. You can check it by using typing ipconfig in command prompt before/after starting the VPN.
Depending upon the number of locations, there are many VPN ls available in the market. I have used SurfShark(Cheap), NordVPN and ExpressVPN(Premium), all three work this way. If you don't have budget issues, you can go for Express, very reliable and higher speeds.
This book includes a checklist for a site audit: https://www.amazon.co.uk/SEO-2019-Actionable-Hands-Including-ebook/dp/B07MCR7HRR/ref=sr_1_1?dchild=1&keywords=seo+2019+andy+williams&qid=1596614436&sr=8-1
I agree. OP should try reading these two books: