Innovate not imitate!

Innovate not imitate!
Interested in the latest Growth hacks?

Welcome to our blog

Interested in the latest Growth hacks?

Welcome to our blog!

We want to help you start/manage and grow your business using innovative strategies and implementation. We have a passion for helping businesses and companies of various sizes see the same success that we have achieved.

Our skillsets are wide and varied, from business strategy, marketing, to online strategy. An increasing number of companies are turning to the internet and online media as a means to maximising their marketing reach and exposure. This is special area of focus for us and we do more than simple SEO strategies.

See our website for more: www.innovatetoaccelerate.com

Tuesday 29 September 2020

Page Authority 2.0: An Update on Testing and Timing

Posted by rjonesx.

One of the most difficult decisions to make in any field is to consciously choose to miss a deadline. Over the last several months, a team of some of the brightest engineers, data scientists, project managers, editors, and marketers have worked towards a release date of the new Page Authority (PA) on September 30, 2020. The new model is exceptional in nearly every way to the current PA, but our last quality control measure revealed an anomaly that we could not ignore.

As a result, we’ve made the tough decision to delay the launch of Page Authority 2.0. So, let me take a moment to retrace our steps as to how we got here, where that leaves us, and how we intend to proceed.

Seeing an old problem with fresh eyes

Historically, Moz has used the same method over and over again to build a Page Authority model (as well as Domain Authority). This model's advantage was its simplicity, but it left much to be desired.

Previous Page Authority models trained against SERPs, trying to predict whether one URL would rank over another, based on a set of link metrics calculated from the Link Explorer backlink index. A key issue with this type of model was that it couldn’t meaningfully address the maximum strength of a particular set of link metrics.

For example, imagine the most powerful URLs on the Internet in terms of links: the homepages of Google, Youtube, Facebook, or the share URLs of followed social network buttons. There are no SERPs that pit these URLs against one another. Instead, these extremely powerful URLs often rank #1 followed by pages with dramatically lower metrics. Imagine if Michael Jordan, Kobe Bryant, and Lebron James each scrimaged one-on-one against high school players. Each would win every time. But we would have great difficulty extrapolating from those results whether Michael Jordan, Kobe Bryant, or Lebron James would win in one-on-one contests against each other.

When tasked with revisiting Domain Authority, we ultimately chose a model with which we had a great deal of experience: the original SERPs training method (although with a number of tweaks). With Page Authority, we decided to go with a different training method altogether by predicting which page would have more total organic traffic. This model presented several promising qualities like being able to compare URLs that don’t occur on the same SERP, but also presented other difficulties, like a page having high link equity but simply being in an infrequently-searched topic area. We addressed many of these concerns, such as enhancing the training set, to account for competitiveness using a non-link metric.

Measuring the quality of the new Page Authority

The results were — and are — very promising.

First, the new model obviously predicted the likelihood that one page would have more valuable organic traffic than another. This was expected, because the new model was directed at this particular goal, while the current Page Authority merely attempted to predict whether one page would rank over another.

Second, we found that the new model predicted whether one page would rank over another better than the previous Page Authority. This was especially pleasing, as it laid to rest many of our concerns that the new model would underperform on old quality controls due to the new training model.

How much better is the new model at predicting SERPs than the current PA? At every interval — all the way down to position 4 vs 5 — the new model tied or out-performs the current model. It never lost.

Everything was looking great. We then started analyzing outliers. I like to call this the “does anything look stupid?” test. Machine learning makes mistakes, just as humans can, but humans tend to make mistakes in a very particular manner. When a human makes a mistake, we often understand exactly why the mistake was made. This isn’t the case for ML, especially Neural Nets; we pulled URLs with high Page Authorities under the new model that happened to have zero organic traffic, and included them in the training set to learn for those errors. We quickly saw bizarre 90+ PAs drop down to much more reasonable 60s and 70s… another win.

We were down to one last test.

The problem with branded search

Some of the most popular keywords on the web are navigational. People search Google for Facebook, Youtube, and even Google itself. These keywords are searched an astronomical number of times relative to other keywords. Subsequently, a handful of highly powerful brands can have an enormous impact on a model that looks at total search volume as part of its core training target.

The last test involves comparing the current Page Authority to the new Page Authority, in order to determine if there are any bizarre outliers (where PA shifted dramatically and without obvious reason). First, let’s look at a simple comparison of the LOG of Linking Root Domains compared to the Page Authority.

Not too shabby. We see a generally positive correlation between Linking Root Domains and Page Authority. But can you spot the oddities? Go ahead and take a minute…

There are two anomalies that stand out in this chart:

  1. There is a curious gap separating the main distribution of URLs and the outliers above and below.
  2. The largest variance for a single score is at PA 99. There are an awful lot of PA 99s with a wide range of Linking Root Domains.

Here is a visualization that will help draw out these anomalies:



The gray spaces between the green and red represent this odd gap between the bulk of the distribution and the outliers. The outliers (in red) tend to clump together, especially above the main distribution. And, of course, we can see the poor distribution at the top of PA 99s.

Bear in mind that these issues are not sufficient to make the new Page Authority model less accurate than the current model. However, upon further examination, we found that the errors the model did produce were significant enough that they could adversely influence the decisions of our customers. It’s better to have a model that is off by a little everywhere (because the adjustments SEOs make are not incredibly fine-tuned) than it is to have a model that is right mostly everywhere but bizarrely wrong in a limited number of cases.

Luckily, we’re fairly confident as to what the problem is. It seems that homepage PAs are disproportionately inflated, and that the likely culprit is the training set. We can’t be certain this is the cause until we complete retraining, but it is a strong lead.

The good news and the bad news

We are in good shape insofar as we have multiple candidate models that outperform the existing Page Authority. We’re at the point of bug squashing, not model building. However, we are not going to ship a new score until we are confident that it will steer our customers in the right direction. We are highly conscientious of the decisions our customers make based on our metrics, not just whether the metrics meet some statistical criteria.

Given all of this, we have decided to delay the launch of Page Authority 2.0. This will give us the necessary time to address these primary concerns and produce a stellar metric. Frustrating? Yes, but also necessary.

As always, we thank you for your patience, and we look forward to producing the best Page Authority metric we have ever released.

Visit the PA Resource Center

Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!



from The Moz Blog https://ift.tt/3ieHe49
via IFTTT

My 8 Best Local SEO Tips for the 2020 Holidays

Posted by MiriamEllis



Image credit: DoSchu

“No place like home for the holidays.” This will be the refrain for the majority of your customers as we reach 2020’s peak shopping season. I can’t think of another year in which it’s been more important for local businesses to plan and implement a seasonal marketing strategy extra early, to connect up with customers who will be traveling less and seeking ways to celebrate at home.

Recently, it’s become trendy in multiple countries to try to capture the old Danish spirit of hygge, which the OED defines as: A quality of coziness and comfortable conviviality that engenders a feeling of contentment or well-being.

While this sometimes-elusive state of being isn’t something you can buy direct from a store, and while some shoppers are still unfamiliar with hygge by name, many will be trying to create it at home this year. Denmark buys more candles than any other nation, and across Scandinavia, fondness for flowers, warming foods, cozy drinks, and time with loved ones characterizes the work of weaving a gentle web of happiness into even the darkest of winters.

Whatever your business can offer to support local shoppers’ aspirations for a safe, comfortable, happy holiday season at home is commendable at the end of a very challenging 2020. I hope these eight local search marketing tips will help you make good connections that serve your customers — and your business — well into the new year.

1) Survey customers now and provide what they want

Reasonably-priced survey software is worth every penny in 2020. For as little as $20/month, your local business can understand exactly how much your customers’ needs have changed this past year by surveying:

  • Which products locals are having trouble locating
  • Which products/services they most want for the holidays
  • Which method of shopping/delivery would be most convenient for them
  • Which hours of operation would be most helpful
  • Which safety measures are must-haves for them to transact with a business
  • Which payment methods are current top choices

Doubtless, you can think of many questions like these to help you glean the most possible insight into local needs. Poll your customer email/text database and keep your surveys on the short side to avoid abandonment.

Don’t have the necessary tools to poll people at-the-ready? Check out Zapier’s roundup of the 10 Best Online Survey Apps in 2020 and craft a concise survey geared to deliver insights into customers’ wishes.

2) Put your company’s whole heart into affinity

If I could gift every local business owner with a mantra to carry them through not just the 2020 holiday shopping season, but into 2021, it would be this:

It’s not enough to have customers discover my brand — I need them to like my brand.

Chances are, you can call to mind some brands of which you’re highly aware but would never shop with because they don’t meet your personal or business standards in some way. You’ve discovered these brands, but you don’t like them. In 2020, you may even have silently or overtly boycotted them.

On the opposite side of this scenario are the local brands you love. I can wax poetic about my local independent grocery store, stocking its shelves with sustainable products from local farmers, flying its Black Lives Matter and LGBTQ+ flags with pride from its storefront, and treating every customer like a cherished neighbor.

For many years, our SEO industry has put great effort into and emphasis on the discovery phase of the consumer journey, but my little country-town grocer has gone leaps and bounds beyond this by demonstrating affinity with the things my household cares about. The owners can consider us lifetime loyal customers for the ways they are going above-and-beyond in terms of empathy, diversity, and care for our community.

I vigorously encourage your business to put customer-brand affinity at the heart of its holiday strategy. Brainstorm how you can make meaningful changes that declare your company’s commitment to being part of the work of positive social change.

3) Be as accessible and communicative as possible

Once you’ve accomplished the above two goals, open the lines of communication about what your brand offers and the people-friendly aspects of how you operate across as many of the following as possible:

  • Website
  • Local business listings
  • Email
  • Social channels
  • Forms
  • Texts/Messaging
  • Phone on-hold marketing
  • Storefront and in-store signage
  • Local news, radio, and TV media

In my 17 years as a local SEO, I can confidently say that local business listings have never been a greater potential asset than they will be this holiday season. Google My Business listings, in particular, are an interface that can answer almost any customer who-what-where-when-why — if your business is managing these properly, whether manually or via software like Moz Local.

Anywhere a customer might be looking for what you offer, be there with accurate and abundant information about identity, location, hours of operation, policies, culture, and offerings. From setting special hours for each of your locations, to embracing Google Posts to microblog holiday content, to ensuring your website and social profiles are publicizing your USP, make your biggest communications effort ever this year.

At the same time, be sure you’re meeting Google’s mobile-friendly standards, and that your website is ADA-compliant so that no customer is left out. Provide a fast, intuitive, and inclusive experience to keep customers engaged.

With the pandemic necessitating social distancing, make the Internet your workhorse for connecting up with and provisioning your community as much as you can.

4) Embrace local e-commerce and product listings

Digital Commerce 360 has done a good job charting the 30%+ rise in online sales in the first half or 2020, largely resulting from the pandemic. The same publication summarizes the collective 19% leap in traffic to North America’s largest retailers. At the local business level, implementing even basic e-commerce function in advance of the holiday season could make a major difference, if you can find the most-desired methods of delivery. These could include:

  • Buy-online, pick up in-store (BOPIS)
  • Buy-online, pick up curbside
  • Buy online for postal delivery
  • Buy online for direct home delivery by in-house or third-party drivers

Here’s an extensive comparison of popular e-commerce solutions, including which ones have free trials, and the e-commerce column of the Moz blog is a free library of expert advice on optimizing digital sales.

Put your products everywhere you can. Don’t forget that this past April, Google surprised everybody by offering free product listings, and that they also recently acquired the Pointy device, which lets you transform scanned barcodes into online inventory pages.

Additionally, in mid-September, Google took their next big product-related step by adding a “nearby” filter to Google Shopping, taking us closer and closer to the search engine becoming a source for real-time local inventory, as I’ve been predicting here in my column for several years.

Implement the public safety protocols that review research from GatherUp shows consumers are demanding, get your inventory onto the web, identify the most convenient ways to get purchases from your storefront into the customer’s hands, and your efforts could pave the way for increased Q4 profits.

5) Reinvent window shopping with QR codes

“How can I do what I want to do?” asked Jennifer Bolin, owner of Clover Toys in Seattle.

What she wanted to do was use her storefront window to sell merchandise to patrons who were no longer able to walk into her store. When a staff member mentioned that you could use a QR code generator like this one to load inventory onto pedestrians’ cell phones, she decided to give it a try.

Just a generation or two ago, many Americans cherished the tradition of going to town or heading downtown to enjoy the lavish holiday window displays crafted by local retailers. The mercantile goal of this form of entertainment was to entice passersby indoors for a shopping spree. It’s time to bring this back in 2020, with the twist of labeling products with QR codes and pairing them with desirable methods of delivery, whether through a drive-up window, curbside, or delivery.

“We’ve even gotten late night sales,” Bolin told me when I spoke with her after my colleague Rob Ousbey pointed out this charming and smart independent retail shop to me.

If your business locations are in good areas for foot traffic, think of how a 24/7 asset like an actionable, goodie-packed window display could boost your sales.

6) Tie in with DIY, and consider kits

With so many customers housebound, anything your business can do to support activities and deliver supplies for domestic merrymaking is worth considering. Can your business tie in with decorating, baking, cooking, crafting, handmade gift-giving, home entertainment, or related themes? If so, create video tutorials, blog posts, GMB posts, social media tips, or other content to engage a local audience.

One complaint I am encountering frequently is that shoppers are feeling tired trying to piecemeal together components from the internet for something they want to make or do. Unsurprisingly, many people are longing for the days when they could leisurely browse local businesses in-person, taking inspiration from their hands-on interaction with merchandise. I think kits could offer a stopgap solution in some cases. If relevant to your business, consider bundling items that could provide everything a household needs to:

  • Prepare a special holiday meal
  • Bake treats
  • Outfit a yard for winter play
  • Trim a tree or decorate a home
  • Build a fire
  • Create a night of fun for children of various age groups
  • Dress appropriately for warmth and safety, based on region
  • Create a handmade gift, craft, or garment
  • Winter prep a home or vehicle
  • Create a complete home spa/health/beauty experience
  • Plant a spring garden

Kits could be a welcome all-in-one resource for many shoppers. Determine whether your brand has the components to offer one.

7) Manage reviews meticulously

Free, near-real-time quality control data from your holiday efforts can most easily be found in your review profiles. Use software like Moz Local to keep a running tally of your incoming new reviews, or assign a staff member at each location of your business to monitor your local business profiles daily for any complaints or questions.

If you can quickly solve problems people cite in their reviews, your chances are good of retaining the customer and demonstrating responsiveness to all your profiles’ visitors. You may even find that reviews turn up additional, unmet local needs your formal survey missed. Acting quickly to fulfill these requests could win you additional business in Q4 and beyond.

8) Highly publicize one extra reason to shop local this year

“72% of respondents...are likely or very likely to continue to shop at independent stores, either locally or online, above larger retailers such as Amazon.” — Bazaarvoice

I highly recommend reading the entire survey of 12,000 global respondents by Bazaarvoice, quantifying how substantially shopping behaviors have changed in 2020. It’s very good news for local business owners that so many customers want to keep transacting with nearby independents, but the Amazon dilemma remains.

Above, we discussed the fatigue that can result from trying to cobble together a bunch of different resources to check everything off a shopping list. This can drive people to online “everything stores”, in the same way that department stores, supermarkets, and malls have historically drawn in shoppers with the promise of convenience.

A question every local brand should do their best to ask and answer in the runup to the holidays is: What’s to prevent my community from simply taking their whole holiday shopping list to Amazon, or Walmart, or Target this year?

Whatever your business can offer to support local shoppers’ aspirations for a safe, comfortable, happy holiday season at home is commendable at the end of a very challenging 2020. I hope these eight local search marketing tips will help you make good connections that serve your customers — and your business — well into the new year.

My completely personal answer to this question is that I want my town’s local business district, with its local flavor and diversity of shops, to still be there after a vaccine is hopefully developed for COVID-19. But that’s just me. Inspiring your customers’ allegiance to keeping your business going might be best supported by publicizing some of the following:

  • The economic, societal, and mental health benefits proven to stem from the presence of small, local businesses in a community.
  • Your philanthropic tie-ins, such as generating a percentage of sales to worthy local causes — there are so many ways to contribute this year.
  • The historic role your business has played in making your community a good place to live, particularly if your brand is an older, well-established one. I hear nostalgia is a strong influencer in 2020, and old images of your community and company through the years could be engaging content.
  • Any recent improvements you’ve made to ensure fast home delivery, whether by postal mail or via local drivers who can get gifts right to people’s doors.
  • Uplifting content that simply makes the day a bit brighter for a shopper. We’re all looking for a little extra support these days to keep our spirits bright.

Be intentional about maximizing local publicity of your “extra reason” to shop with you. Your local newspaper is doubtless running a stream of commentary about the economic picture in your city, and if your special efforts are newsworthy, a few mentions could do you a lot of good.

Don’t underestimate just how reliant people have become on the recommendations of friends, family, and online platforms for sourcing even the basics of life these days. In my own circle, everyone is now regularly telling everyone else where to find items from hand sanitizer to decent potatoes. Networking will be happening around gifts, too, so anything you get noticed for could support extensive word-of-mouth information sharing.

I want to close by thanking you for being in or marketing businesses that will help us all celebrate the many upcoming holidays in our own ways. Your efforts are appreciated, and I’m wishing you a peaceful, profitable, and hyggelig finish to 2020.


Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!



from The Moz Blog https://ift.tt/339k7DN
via IFTTT

Monday 28 September 2020

Synergized search is key to success in the new normal

30-second summary:

  • Given that consumers run billions of searches every day — with Google estimated to process 40,000 per second it’s clear marketers need a smart strategy to cut through the competition.
  • The question is: Will they drive the highest traffic and performance with SEO or PPC?
  • Head of Paid Media at Tug shares insight on how perfectly balancing these two facets can lead to success in the new normal.

Consumer activity online is at an all-time high. So, it’s no surprise many marketers are aiming to make the most of it by hooking their attention early, at the point of search. But deciding how best to do so isn’t necessarily easy.

Given that consumers run billions of searches every day — with Google estimated to process 40,000 per second — it’s clear marketers need a smart strategy to cut through the competition. The question is: will they drive the highest traffic and performance with search engine optimization (SEO) or pay per click (PPC)?

Both have their own advantages and drawbacks. PPC is a quick win, enabling businesses to rapidly reach consumers and boost visibility. But its lead generation power only lasts while the money flows and, depending on campaign scale and scope, those costs can run high. Meanwhile, SEO delivers more lasting rewards and higher click-through rates (CTRs), often for less investment. Yet marketers might have a long wait before organic searches pay off, and may still fall behind dominant digital marketplaces for certain keywords.

Ultimately, the smartest route lies neither one way nor the other, but in a combination of both. Blending PPC and SEO not only generates stronger results but also balances out their respective shortcomings, offering marketers the best chance of success in the new ever-changing normal.

Utilizing a combination of paid and organic search tactics isn’t new – but it’s never been clear how marketers can best do this, or a way to visualize the data for optimization. Leveraging PPC and SEO in conjunction with one another can be challenging, but creating the perfect synergy is possible if marketers focus on the following three factors:

Unify search operations

With consumers spending a quarter of their waking day online, marketers have plenty of chances to spark their interest through search. To outmanoeuvre rivals and capture eyeballs first, brands must make fast yet informed decisions about which approach will produce the ideal outcome.

Achieving this requires holistic insight which, in turn, calls for greater unity. Due to the general view of PPC and SEO as separate entities, teams often operate in silos, but this isolates valuable knowledge around consumer behaviour and the tactics that generate the biggest rewards. Simple as it sounds, removing divisions and encouraging teams to share their insight can significantly improve campaign execution and drive more efficient CPAs.

For example, information from the PPC teams on the best performing keywords and ad copy will help SEO teams to optimize meta descriptions and website content.

Sharing information on what keywords campaigns are covering will also prevent the doubling up of efforts – for example, as organic keyword positions improve, there might be an opportunity to pull back PPC activity and reallocate budget to other keywords to increase the overall coverage. Similarly, updates from the SEO team on keywords that are particularly competitive to rank in top positions are an opportunity for PPC to drive incremental conversions. And, on a more fundamental level, by sharing any new or emerging search terms with each other, both SEO and PPC teams can ensure they are up-to-date and reacting as quickly as possible to opportunities.

Select tech that drives collaboration

The next step is integrated technology. Implementing tools that collate and merge data from multiple sources — including PPC and SEO campaigns — will make collaboration easier. That’s not to mention generating a complete overview of collective search operations, performance, and opportunities for businesses.

A holistic and unified dashboard, for example, can provide visibility of combined search performance against KPIs and competitor activity. This enables PPC and SEO teams to identify where there are opportunities and how strategies can be adjusted to leverage them, without duplicating each other’s efforts. Marketers can understand where organic rankings are high, and competitor activity low, and vice versa, which means they know when to reduce PPC activity, as well as opportunities where it can drive incremental conversions over and above what SEO can deliver.

All of this, however, depends on accuracy and usability. Information needs to be reliable and actionable, which means simply joining up the data dots isn’t enough: in addition to robust cleansing, processing and storage, tools must offer accessible visualization.

Although frequently overlooked, clearly-presented data plays a huge part in enhancing everyday activity. Providing a streamlined picture of keywords and performance data is vital, but to ensure teams can pinpoint prime SERPs, accelerate traffic, and increase conversions, businesses also need tools that allow their teams to quickly find and activate key insights.

Don’t forget human checks

Dialing up tech use, however, does come with a word of warning – no matter how smart platforms may be, they can’t entirely replace human experience and expertise. On their own, sophisticated tools bring a range of benefits that go far beyond translating data into a more cohesive and user-friendly format. The most advanced boast immediate alerts that tell PPC teams where their competitors are bidding — or not — and use artificially intelligent (AI) analysis to deliver a cross-market, sector, and classification perspective on SEO activity.

Human knowledge is still paramount to steering search campaigns in the right direction and picking up on the nuances that machines miss. For instance, problem-solving machines might take the quickest path to objective completion for certain pages or messages, but seasoned search professionals may see the potential for longer-term uses that deliver higher incremental value.

As a result, organizations must avoid the perils of over-reliance on their marketing tools. By persistently applying manual reviews and checking automated conclusions against human knowledge, they can tap the best of tech and people power.

Today’s marketing leaders are grappling with multiple uncertainties, but when it comes to search, the way forward is clear. PPC and SEO are complementary forces; producing deeper insights and higher returns together, as well as minimizing risk. By connecting the two and taking a considered approach to data-driven search strategy, businesses can ensure campaigns are strong enough to succeed in the new normal and take on whatever tomorrow brings.

Asher Gordon is Head of Paid Media at Tug. He leads a multi-disciplined media team who plan, buy, and deliver integrated media plans for a diverse set of clients. With over 10 years experience working across multiple markets and brands at PHD and Wavemaker, Asher works with clients to better their marketing goals and drive their business forward.

The post Synergized search is key to success in the new normal appeared first on Search Engine Watch.



from SEO – Search Engine Watch https://ift.tt/3cJYygb
via IFTTT

Navigating a cookieless future

30-second summary:

  • On September 16, Apple launched iOS 14, which is a major overhaul of the Apple operating system and would require users to authorize information known as (IDFA).
  • This was followed by announcements from Google that they will be following a similar path for Google Chrome, effectively turning off tracking on Safari, which commands 90 percent usage on iPhones, and Chrome, which commands five percent.
  • These moves towards user privacy and marketing compliance are effectively a pivot away from the traditional advertising and search marketing industry, which will impact later players like Facebook to national media agencies like GroupM.
  • More details on how marketers can navigate in a cookieless world.

One of the most impactful changes to internet advertising and media has stayed mostly unspoken in agency and SEO chatter. However, like the switch from a desktop landscape to a mobile landscape, there is no reprieve from the coming cookieless world.

On September 16, Apple launched iOS 14, which is a major overhaul of the Apple operating system and would require users to authorize information known as (IDFA). IDFA is used to track user behavior for advertising.

This was followed by announcements from Google that they will be following a similar path for Google Chrome, effectively turning off tracking on Safari, which commands 90 percent usage on iPhones, and Chrome, which commands five percent.

These moves towards user privacy and marketing compliance are effectively a pivot away from the traditional advertising and search marketing industry, which will impact later players like Facebook to national media agencies like GroupM.

Content created in partnership with SherloQ™, Inc.

National TV advertisers and PPC advertisers are not waiting around

Once again led by the advertising and search category of injury law, due to the highly competitive and expensive nature, we are seeing a couple of key movers.

Smith & Hassler, a nationally recognized personal injury law firm that famously uses Judge Alex Ferrer and William Shatner as TV spokespeople, and Mike Slocumb Law, a firm that is known for its use of celebrity spokespeople and a sometimes outrageous style. They are both first use cases of using Natural Language Understanding (NLU) for content and first-party data extraction, and automatic AI, which assists in marketing automation to Google without the use of cookies.

In both cases, the companies are working with and have implemented  SherloQ™, powered by IBM Watson, to implement cookieless changes and compliance to market their websites.

A recent story from AdWeek quoted Andrew Casale about a cookieless future for publishers, who said it best,

“Publishers haven’t seen a recovery in their CPMs, and similar to Root, believes the focus of online media trading will be publishers’ first-party data as such a method of audience targeting will mean less personal information is traded between (comparatively) anonymous ad-tech players.”

The rapid move towards using first-party data and AI automation will not be limited to a single industry. Privacy is a big selling feature and while Apple has allowed an extension to IDFA, mostly due to the time needed for developers to employ these new frameworks, Apple and Google are not going to wait for the advertising industry’s input.

If your agency or enterprise wants to learn more about how SherloQ™, powered by IBM Watson, can help navigate a cookieless world, please download our white paper to learn more about our framework.

The post Navigating a cookieless future appeared first on Search Engine Watch.



from SEO – Search Engine Watch https://ift.tt/2S54hDR
via IFTTT

How to Detect and Improve Underperforming Content: A Guide to Optimization

Posted by SamuelMangialavori


Content, content, and more content! That’s what SEO is all about nowadays, right? Compared to when I started working in SEO (2014), today, content is consistently one of the most popular topics covered at digital marketing conferences, there are way more tools that focus on content analysis and optimization, and overall it seems to dominate most of SEO news.

Don’t believe me? Here’s a nice Google Trends graph that may change your mind:

Google Trends screenshot for “content marketing” as a topic, set for worldwide interest.

But why is it that content is now dominating the SEO scene? How vital is content for your SEO strategy, actually? And most importantly: how can you be content with your site’s content? Puns aside, this post aims to help you figure out potential causes of your underperforming content and how to improve it.

Why content is key in SEO in 2020

Content is one of the most important factors in SEO. Just by paying close attention to what Google has been communicating to webmasters in the last few years, it’s clear that they’ve put a strong emphasis on “content” as a decisive ranking factor.

For instance, let’s have a look at this post, from August 2019, which talks about Google’s regular updates and what webmasters should focus on:

“Focus on content: pages that drop after a core update don’t have anything wrong to fix. We suggest focusing on ensuring you’re offering the best content you can. That’s what our algorithms seek to reward.”

The article goes on, listing a series of questions that may help webmasters when self-assessing their own content (I strongly recommend reading the entire post).

That said, content alone cannot and should not be enough for a website to rank well, but it is a pretty great starting point!

Underperforming content: theory first

What is underperforming content?

When I say “underperforming content”, I’m referring to content, either on transactional/commercial pages or editorial ones, that does not perform up to its potential. This could be content that either used to attract a good level of organic traffic and now doesn’t, or content that never did generate any organic traffic despite the efforts you might have put in.

Over 90% of content gets no traffic from Google. Ninety bloody percent! This means that nine pages out of 10 are likely not to receive any organic traffic at all — food for thought.

What are the causes of underperforming content?

There could be many reasons why your content is not doing well, but the brutal truth is often simple: in most cases, your content is simply not good enough and does not deserve to rank in the top organic positions.

Having said that, here are the most common reasons why your content may be underperforming: they are in no particular order and I will highlight the most important, in my opinion.

Your content does not match the user intent

Based on my experience, this is a very important thing that even experienced marketers still get wrong. It may be the case that your content is good and relevant to your users, but does not match the intent that Google is showcasing in the SERP for the keywords of focus.

As SEOs, our aim should be to match user intent, which means we first need to understand the what and the who before defining the how. Whose intent we are targeting and what is represented in the SERP will define the strategy we use to get there.

Example: webmasters who hope to rank for a “navigational or informational” keyword with a transactional, page or vice versa.

Your content isn’t in the ideal format Google is prioritizing

Google may be favoring a certain type of format which your content doesn’t conform to, hence it isn’t receiving the expected visibility.

Example: you hope to rank with a text-heavy blog post for a “how to” keyword where Google is prioritizing video content.

Your content is way too “thin” compared to what is ranking

It doesn’t necessarily have to be a matter of content length (there is no proven content length formula out there, trust me) but more relevance and comprehensiveness. It may be the case that your content is simply not as compelling as other sites out there, hence Google prioritizing those over you.

Example: you hope to rank for heavily competitive informational keywords with a 200-words blog post.

Your content isn’t as up-to-date

If your content is very topical, and such a topic heavily depends on information which may change with time, then Google will reward sites that put effort into keeping the content fresh and up-to-date. Apart from search engines themselves, users really care about fresh content — no one wants to read an “SEO guide to improve underperforming content” that was created in 2015!

Example: certain subjects/verticals tend to be more prone to this issue, but generally anything related to regulations/laws/guidelines which tend to change often.

Your content is heavily seasonal or tied to a past event/experience

Self-explanatory: if your content is about something that occurred in the past, generally the interest for that particular subject will gradually decrease over time. There are exceptions, of course (god save the 90s and my fav Netflix show “The Last Dance”), but you get the gist.

Example: topics such as dated events or experiences (Olympics 2016, past editions of Black Friday, and so on) or newsworthy content (2016 US election, Kanye running for president — no wait that is still happening...).

Your tech directives have changed the page’s indexation status

If something happens to your page that makes it fall out of Google’s index. The most common issues could be: unexpected no-index tag, canonical tag, incorrect hreflang tags, page status changes, page removed with Google Search Console’s remove tool, and so on.

Example: after some SEO recommendations, your devs mistakenly put a no-index tag on your page without you realizing.

Your page is victim of duplication or cannibalization

If you happen to cover the same or similar keyword topic with multiple pages, this may trigger duplication and/or cannibalization, which ultimately will result in a loss of organic visibility.

Example: you launch a new service page alongside your current offerings, but the on-page focus (metadata, content, linking structure) isn’t different or unique enough and it ends up cannibalizing your existing visibility.

Your page has been subject to JavaScript changes that make the content hard to index for Google

Let’s not go into a JavaScript (JS) rabbit hole and keep it simple: if some JS stuff is happening on your page and it’s dynamically changing some on-page SEO elements, this may impact how Google indexes your content.

Example: fictitious case where your site goes through a redesign, heavy JS is now happening on your browser and changing a key part of your content that now Google cannot render easily — that is a problem!

Your page has lost visibility following drastic SERP changes

The SERP has changed extensively in the last few years, which means many more new features that are now present weren’t there before. This may cause disruption to previous rankings (hence to your previous CTR), or make your pages fall out of Google’s precious page one.

Also, don’t forget to consider that the competition might have gotten stronger with time, so that could be another reason why you lose significant visibility.

Example: some verticals have been impacted more than others (jobs, flights, and hotels, for instance) where Google’s own snippets and tools are now getting the top of the SERP. If you are as obsessed with SERP chances, and in particular PAA, as I am and want more details, have a read here.

Your content doesn’t have any backlinks

Without going into too much detail on this point — it could be a separate blog post — for very competitive commercial terms, not having any/too few backlinks (and what backlinks represent for your site in Google’s eyes) can hold you back, even if your page content is compelling on its own. This is particularly true for new websites operating in a competitive environment.

Example: for a challenging vertical like fashion, for instance, it is extremely difficult to rank for key head terms without a good amount of quality (and naturally gained) backlinks to support your transactional pages.

How to find the issues affecting your content

We’ve covered the why above, let’s now address the how: how to determine what issue affects your page/content. This part is especially dedicated to a not-too savvy SEO audience (skip this part and go straight to next if you are after the how-to recommendations).

I’ll go through a list of checks that can help you detect the issues listed above.

Technical checks

Google Search Console

Use the URL inspection tool to analyze the status of the page: it can help you answer questions such as:

  • Has my page been crawled? Are we even allowing Google to crawl the page?
  • Has my page been indexed? Are we even allowing Google to index the page?

By assessing the Coverage feature, Google will share information about the crawlability and indexability of the page.


Pay particular attention to the Indexing section, where they mention user-declared canonical vs google-selected canonical. If the two differ, it’s definitely worth investigating the reason, as this means Google isn’t respecting the canonical directives placed on the page — check official resources to learn more about this.

Chrome extensions

I love Chrome extensions and I objectively have way too many on my browser…

Some Chrome extensions can give you lots of info on the indexability status of the page with a simple click, checking things like canonical tags and meta robots tags.

My favorite extensions for this matter are:

JavaScript check

I’ll keep it simple: JavaScript is key in today’s environment as it adds interactivity to a page. By doing so, it may alter some key HTML elements that are very important for SEO. You can easily check how a page would look without JS by using this convenient tool by Onley: WWJD.

Realistically speaking, you need only one of the following tools in order to check whether JavaScript might be a problem for your on-page SEO:

All the above tools are very useful for any type of troubleshooting as they are showcasing the rendered-DOM resources in real-time (different from what the “view-source” of a page looks like).

Once you’ve run the test, click to see the rendered HTML and try and do the following checks:

  • Is the core part of my content visible?
    • Quick way to do so: find a sentence in your content, use the search function or click CTRL + F with that sentence to see if it’s present in the rendered version of the page.
  • Are internal links visible to Google?
    • Quick way to do so: find an internal link on the page, use the search function or click CTRL + F with that sentence to see if it’s present in the rendered version of the page.
  • Can Google access other key elements of the page?
    • Check for things such as headers (example below with a Brainlabs article), products, pagination, reviews, comments, etc.

Intent and SERP analysis

By analyzing the SERP for key terms of focus, you’ll be able to identify a series of questions that relate to your content in relation to intent, competition, and relevance. All major SEO tools nowadays provide you with tons of great information about what the SERP looks like for whatever keyword you’re analyzing.

For the sake of our example, let’s use Ahrefs and the sample keyword below is “evergreen content”:

Based on this example, these are a few things I can notice:

  • This keyword triggers a lot of interesting SERP features (Featured Snippet, Top Stories, People also ask)
  • The top organic spots are owned by very established and authoritative sources (Ahrefs blog, Hubspot, Wordstream etc), which makes this keyword quite difficult to compete for

Here are quick suggestions on what types of checks I recommend:

  • Understand and classify the keyword of analysis, based on the type of results Google is showing in the SERP: any ads showing, or organic snippets? Are the competing pages mainly transactional or informational?
  • Check the quality of the sites that are ranking in page one: indicative metrics that can help you gather insights on the quality of each domain (DA/DR) are helpful, the number of keywords those pages are visible for, the estimated traffic per page, and so on.
  • Do a quick crawl of these pages to bulk check the comprehensiveness of their content and metadata, or manually check some if you prefer that way.

By doing most of these checks, you’ll be able to see if your content is underperforming for any of the reasons previously mentioned:

  • Content not compelling enough compared to what is ranking on page one
  • Content in the wrong format compared to what Google is prioritizing
  • Content is timely or seasonal
  • Content is being overshadowed by SERP features

Duplication and cannibalization issues

Check out my 2019 post on this subject, which goes into a lot more detail. The quick version of the post is below.

Use compelling SEO tools to understand the following:

  • whether, for tracked keywords of interest, two or more ranking URLs have been flip-flopping. That is a clear sign that search engines are confused and cannot “easily decide” on what URL to rank for a certain keyword.
  • whether, for tracked keywords of interest, two or more ranking URLs are appearing at the same time (not necessarily on page one of the SERP). That is a clear signal of duplication/cannibalization.
  • check your SEO visibility by landing page: if different URLs that rank for very similar keyword permutations, chances are there is a risk there.
  • last but not least: do a simple site search for keywords of interest in order to get an initial idea of how many pages (that cover a certain topic) have been indexed by Google. This is an insightful preliminary exercise and also useful to validate your worries.

How to fix underperforming content

We’ve covered the most common cases of underperforming content and how to detect such issues — now let’s talk about ways to fix them.

Below is a list of suggested actions to take when improving your underperforming content, with some very valuable links to other resources (mostly from Moz or Google) that can help you expand on individual concepts.

Make sure your page can be crawled and indexed “properly”

  • Ensure that your page does not fall under any path of blocked resources in Robots.txt
  • Ensure your page is not provided with a no-index meta robots tag or a canonical tag pointing elsewhere (a self-referencing canonical tag is something you may want to consider but not compulsory at all).
  • Check whether other pages have a canonical tag pointing to your URL of focus. Irrelevant or poorly-done canonical tags tend to get ignored by Google — you can check if that is the case in the URL Inspection tool.
  • Ensure your site (not just your page) is free from any non-SEO friendly JavaScript that can alter key on-page elements (such as headers, body content, internal links, etc.).
  • Ensure your page is linked internally on the site and present in your XML sitemap.

Understand search intent

  • Search intent is a fascinating topic in and of itself, and there are a lot of great resources on the subject if you want to delve deeper into it.
  • Put simply, you should always research what the SERP looks like for the topic of interest: by analyzing the SERP and all its features (organic and non), you can get a much better understanding of what search engines are looking for in order to match intent.
  • By auditing the SERP, you should be able to answer the following questions:
    • What type of content is Google favoring here: transactional, navigational, informational?
    • How competitive are the keywords of focus and how authoritative are those competitors ranking highly for them?
    • What content format is Google showcasing in the SERP?
    • How comprehensive should my content be to get a chance to rank in page one?
    • What keywords are used in the competitor’s metadata?
    • What organic features should I consider addressing with my content (things like featured snippets, people also ask, top images, etc.)?
  • Hopefully all the questions above will also give you a realistic view of your chances of ranking on Google’s first page. Don’t be afraid to switch your focus to PPC for some very competitive keywords where your real possibility of organic rankings are slim.

Map your pages against the right keywords

  • This is a necessary step to make sure you have a clear understanding of not only what keywords you want to rank for, but also what keywords you are eligible to rank for.
  • Don’t overdo it and be realistic about your ranking possibilities: mapping your page against several keywords variations, all of which show very different SERPs and intents, is not realistic.
  • My suggestion is to pick two or three primary keyword variations and focus on getting your content as relevant as possible to those terms.

Write great metadata

  • Title tags are still an incredibly important on-page ranking factor, so dedicate the right time when writing unique and keyword-rich titles.
  • Meta descriptions are not a ranking factor anymore, but they still play a part in enticing the user to click on a search result. So from a CTR perspective, they still matter.
  • SEO keyword research is the obvious choice to write compelling metadata, but don’t forget about PPC ad copies — check what PPC ad copies work best for your site and take learnings from them.
  • Don’t change metadata too often, though: make sure you do your homework and give enough time to properly test new metadata, once implemented.

Make the right content amends

  • Based on the intent audit and keyword mapping insights, you’re now ready to work on your actual page content.
  • By now, you’ve done your homework, so you just need to focus on writing great content for the user (and not for Google).
  • Readability is a very important part of a page. Tricks that I’ve learned from colleagues over the years are the following:
    • Read the content out loud and try to objectively assess how interesting it is for your target audience.
    • Make sure to use enough spacing between lines and paragraphs. People’s attention span these days is very short, and chances are people will skim through your content rather than dedicating 100% of their attention to it (I’m sure some of YOU readers are doing it right now!).
    • Make sure your tone of voice and language match your target audience (if you can write things in plain English vs. highly technical jargon, do so and don’t over-complicate your life).
  • Make sure you’ve thought about all internal linking possibilities across the site. Not only for the same type of page (transactional page to transactional page, for instance) but also across different types (transactional page to video/blog post, if that helps people make a decision, for example).
  • Optional step: once everything is ready, request indexing of your page in Google Search Console with the URL inspection tool.

Final thoughts

Underperforming content is a very common issue and should not take you by surprise, especially considering that content is considered among (if not the) most important ranking factors in 2020. With the right tools and process in place, solving this issue is something everyone can learn: SEO is not black magic, the answer tends to be logical.

First, understand the cause(s) for your underperforming content. Once you’re certain you’re compliant with Google’s technical guidelines, move on to determining what intent you’re trying to satisfy. Your research on intent should be comprehensive: this is what’s going to decide what changes you’ll need to make to your content. At that point, you’ll be ready to make the necessary SEO and content changes to best match your findings.

I hope this article is useful! Feel free to chat about any questions you may have in the comments or via Twitter or LinkedIn.


To help us serve you better, please consider taking the 2020 Moz Blog Reader Survey, which asks about who you are, what challenges you face, and what you'd like to see more of on the Moz Blog.

Take the Survey

Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!



from The Moz Blog https://ift.tt/3i5POSN
via IFTTT

Friday 25 September 2020

Top 10 Tools to facilitate web development

30-second summary:

  • Web development is very essential for guiding you past the traditional methods of website development and approaching to other more unique and efficient prospects in the IT field.
  • Building an app or website is real but sometimes can be better facilitated through the testing and implementation phases of web tools.
  • Entering the new age of the digital world is just impossible without proper deployment, scaling, and database management. And this where the best web development weapons are on standby.

Web development tools are necessary to adapt and integrate in the development environments since they help you create a webpage, gather metrics, discover layouts and designs. Web development is all about evolving and stepping up the game you are currently very proficient in. Along with that it doesn’t demand much money,  require enormous sources or expect big money. The only thing development tools are incorporating is your boosted ultimate website performance. What is simply left is finding the right assistant on this journey. Here are the top 10 we have picked for you:

1. Chrome DevTool

If you are searching for a web assistant that would ensure your safety and optimize development, Chrome DevTool is sincerely recommended in this case. Taking into account that this is the most popular cross-platform browser of all, you should be certain that such a Chromium-based development technology will allow you to inspect, debug, and guide you through the browser.

While varied development tools are primarily focused on setting up websites, creating online stores, and launching blogs, Chrome DevTool is highly helpful in inspecting and tweaking. Although it is a bit hard to navigate through so many tabs and functions, Chrome DevTool offers you numerous panels for boosting productivity, analyzing load performance, and enhancing better user-friendly experience:

  • Elements panel
  • Console panel
  • Sources panel
  • Network panel
  • Performance panel
  • Memory panel
  • Application panel
  • Security panel

Such a division gives you a more fixed structure on elements and commands that you can apply depending on the existing page you are willing to diagnose. In order to access Chrome DevTools, you won’t even need to pay for anything, as it is built into the browser itself. Here is how you can open it.

2. VironIT

Good software development company’s support can facilitate your user experience with its comprehensive and all-inclusive approach even more. VironIT offers you a list of the new age web development tools and services to create a highly adaptable website or attractive web application.

If you wonder what kind of development tools are combined in one software, it might actually vary from AR&VR development (augmented and virtual reality) to IoT Software development (data visualization and analysis). Such a useful package of cross-platform tools in one place makes it easier for you to write, maintain and promote better development and react to possible vulnerabilities. In addition, it guarantees you rapid functionality of apps, programs, websites and its well-organized software designs.

There is nothing more efficient than a versatile service providing the fast and most secure web frameworks. You share the main goal.

3. React JS

React is a JavaScript library for building user interfaces and reusable UI components. Why do many developers use this web development tool is mainly because of the ease of breaking a developing application into several simple parts. Moreover, the style of application code derived with the help of React components can be used for any program’s interface which significantly speeds the process up.

Obviously, if you are willing to learn ReactJS, you need to know how to work with JavaScript, HTML5, and CSS. The solid knowledge of any of these markup languages makes the optimization journey with ReactJS more helpful and enjoyable:

  • It is faster because it performs optimization while drawing up code to JavaScript.
  • It is also error-free since all technical mistakes can be spotted during compilation.
  • It is easier to write templates, especially when you are familiar with HTML.
  • It is entirely automated since it can handle the JavaScript code, which is suitable for many browsers.

4. Github

You have no clue what the new age of web development means if you don’t try out GitHub. This web platform enables you not only to manage the web applications by coordinating the tools and its usability, but also to deploy regular teams and campaigns to the completion of these same applications or projects. GitHub accesses you a lightweight, branch-based workflow that helps you experiment with taking over new tasks and managing the existing workflow.

  • It helps you make changes, reviews without affecting other involved branches.
  • It offers a smooth documentation process that often complements the quality coding process.
  • It can modify pages’ content and style remotely via the web or locally on your computer

Why GitHub? The suggested privileges of breaking down barriers within a team when the project is implemented and keeping everyone focused on having the best performance. In the same way, it can let your business flow run smoothly and consistently at any point in time.

5. Novi Builder

You can’t even imagine what excellent sophisticated features you have been holding back when you were not a member of the Novi Builder community. This easy-to-use development tool offers you facilities for HTML drag and drop builder, access to all existing and future templates including free of charge themes. Also, it provides you with 20 must-have plugins and elements, such as Social Plugins, Contact form, Sliders, Countdown, Google Maps, and more.

The functionality and designs are exactly built for beginners developers who want to evolve and rise up on their IT working ladder. It is a great chance for deepening your understanding of website creation and ensuring solid knowledge of application development.

Along with the above-mentioned features, here is what else Novi Builder has to offer:

  • It offers advanced code editor support.
  • It supercharges your projects just from scratch
  • It can install an app and make it live in several hours at low cost
  • It empowers and modifies your existing HTML website with additional application integration facilities

6. BugMuncher

The BugMuncher’s list of integrations includes GitHub, Jira, Trello making it more accessible and easier to edit packages and manifest files. The toolkit enables you to draw the problems captured on your website and send it across as a screenshot. Generally, BugMuncher teaches developers how to make visual highlights and get immediate feedback with the most accurate details. In addition, this feedback comes not just from random visitors but your current customers when they interact with your service and share the reports of JavaScript and AJAX errors.

Unlike the rest of the feedback tools, this one doesn’t misuse or distort gathered information. It does exactly the opposite by testing development phases and saving your existing project from bugs. If you are quite hesitant it takes too long to collect BugMuncher feedback, you should remember that it is not about your colleague’s professional suggestion, a customer as a prime source of feedback. They can deliver you the most constructive criticism that would help you hone your business’ website or app development.

7. Bootstrap

Bootstrap is low-key the most popular HTML, CSS, and JavaScript framework for developing responsive, mobile-first websites. Its open-source toolkit stores different templates and designs for creating common user interface components. These elements usually include forms, buttons, navigations, dropdowns, alerts, modals, tabs, accordions, carousels, tooltips, and so on. Basically, it allows you to build mobile-friendly websites with a smart drag and drop facility.

Moreover, ready-made blocks of code and flexible layouts needed for HTML, CSS can be easily customized and used in accordance with your project’s requirements. The greatest Bootstrap’s features of all are:

  • Extensive list of components
  • Solid guarantee of the project’s responsiveness
  • Useful icon library

8. Notepad++

Notepad++ is your great assistant in text editing, syntax highlighting, plugin maintaining and this list of handy features can actually go on. Although it is available on Windows only, it can still be supported by GitHub, a primary distribution party.

Additionally, Notepad++ supports 27 programming languages. If you go to the Binary Translations page you can easily choose the language you need. Alternatively, you can open User Language Define System to help you define the definition of keywords or regular expressions. Moreover, if you are advanced at coding in languages like C++, Batch, and HTML, Notepad++ is just an ideal service for you. Imagine that:

  • You can find every instance of a word and perform actions on it.
  • You can access the replace function and get a change reference for each word.
  • You can also save time by typing the keyboard shortcut and skipping the appeared code next time or previous time.

Doesn’t it sound amazing to apply a word-search and text-editor tool like this with just a quick installation and some default language setting? You should definitely try it out.

9. Source tree

We wouldn’t bet on the wrong horse if we said that most of you are intermediate or even advanced developers who’d rather focus on coding than interacting with the Git repositories. SourceTree is a free web development tool available for Mac and Windows and perfect for database visualization and management. However, Source tree equally claims to serve both powerful experts, as well as absolute beginners. Yet, understanding the interface specs is easier than you can even think:

  • Bookmark windows represent database you are aligned with
  • Toolbar lets you take action without any command line
  • The sidebar is where you access the information about your current project
  • Repository history and code diffs share control to monitor previous changes

From now any coding changes will be demonstrated in real-time and with the aggregated view on all of your work.

10. Lambda Test

Certainly, before displaying your website to the users’ eyes, as a developer you are in charge of rendering it over different browsers. Depending on the coding language you use (like HTML, CSS  & JavaScript) they should be tested beforehand. Lambda Test is your main guide to ensure that your web application is adaptable for any browser even after you push code changes. Such cross-browser test automation offered by Lambda Test has a very simple and easy to understand interface. Even more, better addition is a vast collection of integration facilities (screenshots screen recording, plugins, and so on). All in all, this unique rendering mechanism is less complex, time-consuming, and instead much cheaper and faster in use.

Conclusion

Now, you are completely forearmed to successfully lead the war of web development. Getting an application done or an advanced code supported has never been easier. Try these top tools out and you will not regret it.

The post Top 10 Tools to facilitate web development appeared first on Search Engine Watch.



from SEO – Search Engine Watch https://ift.tt/342y9WO
via IFTTT

Speed and performance of Web dev, SEO, and marketing agencies websites

30-second summary:

  • The research shows most agencies failed when it comes to the performance of their website.
  • Search engine ranking is a multi-factor game, and performance, while it matters for many reasons, is just one piece in this puzzle.
  • Nebojsa Radakovic shares insights.

Ever since Google announced that page speed would be a ranking factor in its mobile-first index in 2018, the need for speed became one of the most important aspects of web dev trait. A lot of businesses jumped onto the speed train. 

Sure enough, one year later, Google reported that sites are faster, and abandonment rates are down since making page speed a ranking factor.

With performance being one of the top-selling points of a modern-day web dev architecture Jamstack that we are so into, it was only natural to take a deep dive into the industries that tackle website performance and see how we stand against our peers.

TL;DR: Key findings

Don’t have the time to read through the research? Here are the key findings:

  • 27% of websites from our 20K sample still run on HTTP
  • 65.7% of the websites are built with WordPress
  • Only 2.7% of websites have good performance scores
  • 2.9% of websites provide good user experience to their users, ie Largest Contentful Paint (LCP) occurs within 2.5 seconds of when the page first starts loading

What data was I interested in, and why?

Lighthouse performance metrics. There are a couple of popular speed testing tools, but most people use Lighthouse. While it may not be perfect because it provides a mix of both lab and field data about a page, I’ve used Pagespeed Insights API as described in James McNulty UpBuild post here, although updated to show core web vitals.

CMS. WordPress or not. 37% of all websites are powered by WordPress. Being the most popular web dev solution, it would be interesting to see and compare different solutions in terms of speed and performance.

Where did I get my URLs from?

Gathering URLs is a time-consuming work. But I managed to get 20k URLs (20397 URLs to be exact). I’ve cross-referenced results I got from scraping the first-page organic results of a set of keywords (like SEO agency, web dev agency, etc.), results I got by using tools such as Phantombuster to scrap review websites, and results I got from hiring virtual assistants on Upwork and Fiver.

There are a couple of issues I had to take care of first. Amazingly 27% of websites from my 20K sample still run on HTTP. That’s not good at all. On top of that, I had a bunch of URLs coming up with NET::ERR_CERT_DATE_INVALID error message in Chrome. Once those were taken care of, I ended up having results 13945 URLs instead of 20K.

The data

Popular CMS- Speed and performance research

Of course, the most popular CMS is WordPress, with 65.7% of websites from my sample using it. For 18.8%, I was not able to detect any CMS. 2.58% run on Squarespace, 1.6% are built with Drupal, 1.41% are on Wix, and so on.

The results should not come as a surprise given that WordPress powers 37% of all the websites on the Internet or 63.6% of all the websites with known CMS.

Performance scores – How scores are color-coded by Google

The metrics scores and the perf score are colored according to these ranges:

  • 0 to 49 (Red): Poor
  • 50 to 89 (Orange): Needs Improvement
  • 90 to 100 (Green): Good

You can read more about it here.

As far as the performance scores for all websites are concerned, 77.1% of the websites are in the poor range, which means there is a lot of room for improvement.

URL performance - Speed and performance research

Pretty much the same story when we check only WordPress websites, 83.9% are in the poor performance range.

Performance WordPress - Speed and performance research

Core Web Vitals

By now, you probably are well aware of Core Web Vitals. Their importance is twofold:

  • Google considers them essential in a webpage’s overall user experience, and understanding them can help you improve the quality of experience you are delivering to your users,
  • Google plans to make a page experience an official Google ranking factor with Core Web Vitals being an essential part of it.

The current set for Core Web Vitals focuses on three aspects of the user experience: loading (described with Largest Contentful Paint (LCP) metric), interactivity (described with First Input Delay (FID) metric), and visual stability (described with Cumulative Layout Shift (CLS) metric).

For this research, numbers follow the performance scores. For example, check out the Largest Contentful Paint (LCP) results.

Largest-Contentful-Paint-(LCP): Speed and performance analysis

Conclusion

Being that I’ve tested only 20k URLs (actually 13945), let’s not generalize conclusions. However, the general ‘feel’ is that the ones required to think of speed and performance failed the test.

Performance, while it matters for many reasons, is not and should not be the end goal. It depends not only on the tech used but also ‘features’ you’ll have on a website, which pretty much depends on the industry/theme your website is in. And balancing performance and functionality successfully depends on the value a feature brings to your business versus the reduction in speed that results.

The thing is, whatever tech you use, you can end up with good scores (some easier than others). The real question is, how important are the scores for your client, their business, and their audience?

Nebojsa Radakovic is an SEO wiz with 20 years of experience. He is also an extreme sports enthusiast. He can be found on Twitter @CookieDuster_N.

The post Speed and performance of Web dev, SEO, and marketing agencies websites appeared first on Search Engine Watch.



from SEO – Search Engine Watch https://ift.tt/34ewjTd
via IFTTT

Accessible Machine Learning for SEOs — Whiteboard Friday

Posted by BritneyMuller

Machine learning — a branch of artificial intelligence that studies the automatic improvement of computer algorithms — might seem far outside the scope of your SEO work. MozCon speaker (and all-around SEO genius) Britney Muller is here with a special edition of Whiteboard Friday to tell you why that's not true, and to go through a few steps to get you started. 

To see more on machine learning from Britney and our other MozCon 2020 speakers, check out this year's video bundle. 

Get my MozCon 2020 video bundle

Accessible Machine Learning

Click on the whiteboard image above to open a high resolution version in a new tab!

Video Transcription

Hey, Moz fans. Welcome to this special edition of Whiteboard Friday. Today we are taking a sneak peek at what I spoke about at MozCon 2020, where I made machine learning accessible to SEOs everywhere.

This is so, so exciting because it is readily at your fingertips today, and I'm going to show you exactly how to get started. 

So to kick things off, I learned about this weird concept called brood parasites this summer, and it's fascinating. It's basically where one animal tricks another animal of the same species to raise its young.

It's fascinating, and the more I learned about it, the more I realized: oh my gosh, I'm sort of like a brood parasite when it comes to programming and machine learning! I latch on and find these great models that do all the work — all of the raising — and I put in my data and my ideas, and it does things for me.

So we are going to use this concept to our advantage. In fact, I have been able to teach my dad most of these models that, again, are readily available to you today within a tool called Colab. Let me just walk you through what that looks like. 

Models to get you started

So to get started, if you want to start warming up right now, just start practicing clicking "Shift" and then click "Enter".

Just start practicing that right now. It's half the battle. You're about to be firing up some really cool models. 



All right. What are some examples of that? What does that look like? So some of the models you can play with today are things like DeOldify, which is where you repair and colorize old photos. It's really, really fun. 

Another one is a text generator. I created one with GTP-2 — super silly, it's this excuse generator. You can manipulate it and make it do different things for you. 

There's also a really, really great forecasting model, where you basically put in a chunk of time series data and it predicts what the future might have in store. It's really, really powerful and fun.

You can summarize text, which is really valuable. Think about meta descriptions, all that good stuff. 

You can also automate keyword research grouping, which I'll show you here in a second. 

You can do really powerful internal link analysis, set up a notebook for that.

Perhaps one of the most powerful things is you can extract entities and categories as Google perceives them. It's one of my favorite APIs. It's through Google's NLP API. I pull it into a notebook, and you basically put the URLs you want to extract this information from and you can compare how your URL compares to competitors.

It's really, really valuable, fun stuff. So most importantly, you cannot break any of this. Do not be intimidated by any of the code whatsoever. Lots of seasoned developers don't know what's happening in some of those code blocks. It's okay.

Using Colab

We get to play in this environment. It's hosted in Google Drive, and so there's no fear of this breaking anything on your computer or with your data or anything. So just get ready to dive in with me. Please, it's going to be so much fun. Okay, so like I said, this is through a free tool called Colab. So you know how Google basically took Excel and made Google Sheets?

They did the same thing with what's known as Jupyter Notebooks. So these were locally on computers. It's one of the most popular notebook environments. But it requires some setup, and it can be somewhat clunky. It gets confused with different versions and yada, yada. Google put that into the cloud and is now calling it Colab. It's unbelievably powerful.

So, again, it's free. It's available to you right now if you want to open it up in a new tab. There is zero setup. Google also gives you access to free GPU and TPU computing, which is great. It has a 12-hour runtime. 

Some cons is that you can hit limits. So I hit the limits, and now I'm paying $9.99 a month for the Pro version and I've had no problems.

Again, I'm not affiliated with this whatsoever. I'm just super passionate about it, and the fact that they offer you a free version is so exciting. I've already seen a lot of people get started in this. It's also something to note that it's probably not as secure or robust as Google's Enterprise solution. So if you're doing this for a large company or you're getting really serious about this, you should probably check out some other options. But if you're just kind of dabbling and want to explore and have fun, let's keep this party going. 

Using pandas

All right. So again, this is basically a cloud hosted notebook environment. So one thing that I want to really focus on here, because I think it's the most valuable for SEOs, is this library known as "pandas".

Pandas is a data frame library, where you basically run one — or two — lines of code. You can choose your file from your local computer, so I usually just upload CSVs. This silly example is one that I really did run with Google Search Console data.

So you run this in a notebook. Again, I'm sharing this entire notebook with you today. So if you just go to it and you do this, it brings you through the cells. It's not as intimidating as it looks. So if you just click into that first cell, even if it's just that text cell, "Shift + Enter", it will bring you through the notebook. 


So once you get past and once you fire up this chunk of code right here, upload your CSV. Then once you upload it, you are going to name your data frame. 


So these are the only two cells you need to really change or do anything with if you want. Well, you need to. 

So we are uploading your file, and then we are grabbing that file name. In this case, mine was just "gsc-example.csv". Again, once you upload it, you will see the name in that output here. So you just put that within this code block, run this, and then you can do some really easy lines of code to check to make sure that your data is in there.


So one of the first ones that most people do is "df". This is your data frame that you named with your file right here. So you just do "df.head()". This shows you the first five rows of your data frame. You can also do "df.tail()", and it shows you the last five rows of your data frame.

You can even put in a number in here to modify how many rows you want to explore. So maybe you do "df.head(30)", and then you see the first 30 rows. It's that easy just to get it in there and to see it. Now comes the really fun stuff, and this is just tip of the iceberg.

So you can run this really, really cool code cell here to create a filterable table. What's powerful about this, especially with your Google Search Console data, is you can easily extract and explore keywords that have high click-through rate and a low ranking in search. It's one of my favorite ways to explore keyword opportunities for clients, and it couldn't be easier.

So check that out. This is kind of the money part right here. 

If you're doing keyword research, which can take a lot, right, you're trying to bucket keywords, you're trying to organize topics and all that good stuff, you can instantly create a new column with pandas with branded keyword terms.

So just to walk you through this, we're going "df["Branded"]". This is the name of the new column we're going to create. We have this query string "contains," and this is just regex, ("moz|rand|ose"). So any keywords that contain one of those words gets in the "Branded" column a "True".

So now that makes filtering and exploring that so much faster. You can even do this in ways where you can create an entirely different data frame table. So sometimes if you have lots and lots of data, you can use the other cell in that example. All of these examples will be in the notebook.

You can use that and export your keywords into buckets like that, and there's no stall time. Things don't freeze up like Excel. You can account for misspellings and all sorts of good stuff so, so easily with regular expressions. So super, super cool.

Conclusion

Again, this is just tip of the iceberg, my friends. I am most excited to sort of plant this seed within all of you so that you guys can come back and teach me what you've been able to accomplish. I think we have so much more to explore in this space. It is going to be so much fun. If you get a kick out of this and you want to continue exploring different models, different programs within Colab, I highly suggest you download the Colab Chrome extension.

It just makes opening up the notebook so much easier. You can save a copy to your drive and play with it all you want. It's so much fun. I hope this kind of sparked some inspiration in some of you, and I am so excited to hear what all of you think and create. I really appreciate you watching.

So thank you so much. I will see you all next time. Bye.

Video transcription by Speechpad.com


Ready for more?

You'll uncover even more SEO goodness in the MozCon 2020 video bundle. At this year's special low price of $129, this is invaluable content you can access again and again throughout the year to inspire and ignite your SEO strategy:

  • 21 full-length videos from some of the brightest minds in digital marketing
  • Instant downloads and streaming to your computer, tablet, or mobile device
  • Downloadable slide decks for presentations

Get my MozCon 2020 video bundle


Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!



from The Moz Blog https://ift.tt/2G5IskM
via IFTTT

Social Media Today