Category Archives: News-Events

How To Upload Videos To YouTube for the Right sizes – Step by Step


How To Upload Videos To Youtube  – Video Display Resolution

Using an improper display size or aspect ratio is a common mistake you will discover in web videos. In this article we are going to be talking about how you can get better at uploading videos for your website.

Problem 1

A typical thing that happens when you visit someone’s website or check out some YouTube videos, is that you may notice black bars that appear on the left and right side of the video, as illustrated in the image below.

The image above shows an example of someone who recorded their video using a 4:3 aspect ratio, which is the ratio of the video’s width to video’s height. This means that for every 4 pixels that appear horizontally, there are 3 pixels that appear vertically.

The problem is that platforms such as YouTube do not use a 4:3 aspect ratio, instead they use 16:9 aspect ratio. Hence, when you upload a video with a 4:3 aspect ratio on YouTube, it’s like trying to fit a square into a rectangle.

Hence, when someone uploads a video that is the wrong size, YouTube centers the video in the middle of the screen and then fill the empty parts with black bars (where the red arrows point) on the left and right side of the video.

You’ve probably come across such a case, and it is a very simple problem to avoid. In future to avoid this kind of problem, you must use a 16:9 aspect ratio! When you record almost

here is the PDF version to print Click Here 

anything that is going to be used for web content, particularly YouTube content, make sure you are recording at an aspect ratio of 16:9.

Whether it’s your screen recorder software or video camera make sure its aspect ratio is set to a recording size that is compatible with an aspect ratio of 16:9. The following is a list of 5 recording ratios compatible with this ratio:

  • 426 x 240 (240p)
  • 640 x 360 (360p)
  • 854 x 720 (480p)
  • 1280 x 720 (720p)
  • 1920 x 1080 (1080p)

All these sizes will perfectly fit into YouTube’s default 16:9 video player because each of these ratios are just bigger multiples of 16 and 9. It is important to be very conversant with these sizes if you want to get good at video creation.

Problem 2

The second problem with the video is that the text in the video appears to be very small. What usually happens when you upload a video on YouTube is that it creates five different file sizes available: 240p, 360p, 480p, 720p and 1080p.

If you click on the little gear at the bottom of the YouTube video, you can view all the five different sizes available as long as the original video size was big enough. This is illustrated in the image below.

If the original video was not big enough, YouTube just creates as many possible sizes as they can. 640 X 360 is the default YouTube player size. It is a small window to work with, but YouTube keeps this window small because videos are often usually very large files. Considering that it is among the top 3 most visited websites in the world, they have to deliver a lot of data per second and most users are usually connected to a low bandwidth internet connection.

Therefore, if YouTube were to increase its video player file size, then this would result to increased waiting time. Waiting time increases as video file size increases, as shown in the image below.

A large video file size means that the users would have to wait longer for the video to queue up. For every second your users will have to wait, you will lose a number of customers. It is thus important to note that the vast majority of users will watch your videos through the default 640 x 360 YouTube player size.

Since this article is mainly targeted towards internet marketers and website owners, we are going to focus on the 3 video sizes that are suitable for these users.

Out of the 5 video sizes, we are going to eliminate 2 video sizes:

  • The 426 x 240 (240p) aspect ratio, because this setting is just too small. They are not of enough high quality to qualify their use in this day and age.
  • The 1920 x 1080 (1080p) aspect ratio, because it is mostly for high resolution videos and the majority of users will not have the patience to wait for 1080p videos to finish loading.

It is important to note that 640 x 360 (360p) is YouTube’s video player’s file size and the majority of our users will view your videos through this file size. Therefore, if your goal is to lure YouTube views, it is important to look acceptable at this size.

As internet marketers and website owners, it is crucial to maintain our videos at the following video file sizes:

  • 640 x 360 (360p)
  • 854 x 720 (480p)
  • 1280 x 720 (720p)

Recording Softwares

There are many video recording software out there, but here is a list of the top 3 recording softwares that will give you the best results:

  1. The Camtasia Recorder (not free)
  2. Freez Screen Video Capture (free and operates only on windows)
  3. (free and operates on mac, however, it has limited features)

Take away Message.

No matter which video recording software you decide to use, the message you should carry home is that you should always remember to type in the correct values of aspect ratios when you are recording.

Almost all video recording softwares will have little boxes where you can type in the width and height of the area you are about to record. Before clicking the record button, you have to make sure you type in one of the suggested values with a 16:9 aspect ratio, which are YouTube compliant sizes.


If you liked this article share it with friends and also leave your thoughts below.


Again here is the Form where you guys can Print it  Click here


Footer Link Optimization for Search Engines |Interrelated Sitewide Footer Links

According to online marketing expert and SEO guru, Mike Grehan, on a topic related to search engine rankings, he explained that sites ranked high on Google tend to maintain their current positions. He applied the term “the rich get richer” to describe this situation during an interview with the co-founder of Emperical proof, Chris Lake.sitewide-links-footer

In this article, we are going to explain why this is the case.

One contributing factor is discoverability of the site.

A writer conducting a search on Google in most cases is likely to click on the top result before trying to get information from the other bottom results. In the event that the webpage correctly and fully answers the researcher’s questions, it is possible to gain a citation in that writer’s article. Such a citation has the impact of strengthening that webpage’s position.

In terms of search positions, the rich get richer. This is with reference to an exhaustive post authored by Glen Allsopp (

The rich in this context are major publishing groups.

One contributing factor is discoverability of the site.

A writer conducting a search on Google in most cases is likely to click on the top result before trying to get information from the other bottom results. In the event that the webpage correctly and fully answers the researcher’s questions, it is possible to gain a citation in that writer’s article. Such a citation has the impact of strengthening that webpage’s position.

In terms of search positions, the rich get richer. This is with reference to an exhaustive post authored by Glen Allsopp (

The rich in this context are major publishing groups.

The means by which they are achieving higher search engine rankings is by cross-linking to existent and fresh websites, from body copy and footers which are “constantly altered”.

This strategy is not horrifically wrong, but there is some risk in making adjustments to links often, particularly when the anchor is not something related to the site’s brand name.

As Glen pointed out, for anyone who has implemented search engine optimization techniques and strategies over a long period of time, getting many sitewide links within a short timeframe should raise a red flag to search engines’ webspam team.

There are interesting conclusions made from Glen’s posts that show that Google seems to not only reward this kind of conduct but also puts up with it.

Hearst is one of the sites that were found to be taking part in this kind of behavior. This site was discovered to be linked to a freshly launched website, BestProducts, from its listing of authority websites. This portfolio includes the likes of Bazaar and Cosmopolitan.

The use of this strategy rewarded the new site in a striking way

Have fun in footerland

Below are screenshots. Note that in March, the anchor text was ‘Style Reviews’.

The second screenshot below was taken later. The link text changed toBeauty Reviews’. Another thing to note is that the link placement was also altered.

Sidenote: here is the HTML code to use it wisely on footer link

If we assume that these links are dofollow, they have the potential to attract the feared manual penalty for some site owners.

Clearly, these links were changed by design rather than by accident.

The question to ask ourselves is; did the applied strategy work?


According to Glen, the strategy worked astonishingly well. He estimates that Best Products attracted not less than 600,000 organic referrals from Google in April 2016.

Below is a list of positions that Best Products has bragged about in just over half a year, since it was launched.

Their Top Keywords According to SEMRush

Some of those incredible rankings they’ve achieved include:

  • hairstyles: 11th (450,0000 searches per month)
  • short hairstyles: 7th (301,000 searches per month)
  • best wireless earbuds: 1st (22,200 searches per month)
  • short haircuts: 9th (301,000 searches per month)
  • best running shoes for women: 1st (18,100 searches per month)
  • bluetooth speakers: 11th (165,000 searches per month)
  • lighted makeup mirror: 1st (14,800 searches per month)
  • best makeup brushes: 1st (14,800 searches per month)
  • haircuts: 7th (165,000 searches per month)
  • short haircuts for women: 6th (110,000 searches per month)

Outstanding, right? It’s is easy to spot some very broad terms there.

Glen considers the following 16 companies, as well as the brands they own, dominate Google’s results.

There is a possibility that where major brands own hundreds of sub-brands, there is application of similar tactics.

Is This Conduct Acceptable?

The question here is not whether Hearst and its equals are outsmarting Google with their clear site wide link tactic, nor whether the tactic is reliable in the long term. It is solely to figure out whether Google comprehends related entities.

Can Google really identify that these sites are linked together by the same parent company? And is this a tactic that Google deems acceptable?

We know for a fact that if a site was flagged for a manual action, the webspam team can definitely spot that one site is linked to another. Does this mean that Google turns a blind eye to this kind of conduct?

Back in 2014, this is what Matt Cutts had to say about related sites:

“If you have 50 different sites, I wouldn’t link to all 50 sites down in the footer of your website, because that can start to look pretty spammy to users. Instead you might just link to no more than three or four or five down in the footer, that sort of thing, or have a link to a global page, and the global page can talk about all the different versions and country versions of your website.”

“If you’ve got stuff that is all on one area, like .com, and you’ve got 50 or 100 different websites, that is something where I’d be really a lot more careful about linking them together.”

“And that’s the sort of thing where I wouldn’t be surprised if we don’t want to necessarily treat the links between those different websites exactly the same as we would treat them as editorial votes from some other website.”

It is important to note that Matt talks about links to other sites, not “links with descriptive and ever-changing anchor text”, which are two different things.

Drop Hub Pages and Instead Launch Hub Sites

Internal linking is more powerful when there is a distinct strategy in place. You have to study the systematic and usual vocabulary beforehand. The path created should reflect the main business goals and should be suited for your visitors.

Even with all that in mind, it still takes a solid authority site to pull this off. Thus…the rich get richer.

In the event that the strategy works in the long run, there are expectations to see a lot of niche sites achieving dramatic search engine rankings after being launched, as a result of cross-site linking with major publishing groups.

Takeaway Message

It is apparent that sitewide footer linking approach is a strategy that is working and we have not heard the last of its story. We can conclude that the power of links in setting a site on the map cannot be underestimated.

If you liked this post share it with friends and also leave your thoughts below.





There have been plenty of changes happening and others still expected to happen in the ever evolving web industry. Apart from just updating you on the latest trends, I will also be informing you on some of the steps you should be taking to keep up with these changing trends.

Okay, here we go…

  1. Your Website Has to Load FAST

In a generation where almost everyone is thirsty for information, no one has the patience to wait for a slow website to finish loading. Only fast loading websites will prove beneficial to users.

In response to this, we are seeing implementation of new features such as Accelerated Mobile Pages and Facebook Instant Articles (IA). The purpose of these developments is to allow websites to load faster on Social media platforms as well as on Google.

Despite the fact that users are getting instant access to information, it is creating a problem for marketers because they are experiencing a decline in the amount of traffic to their websites.

These websites still get acknowledgement for visits in analytics, but the engagement metrics take a huge hit. This means it becomes difficult to measure the effect of their content.

Take-home message:

These developments have brought a huge influence on user experience and soon marketers and bloggers who will not have embraced these changes will start to suffer immensely.

In future, users will want to focus on instant articles and won’t want to visit websites as it will be a waste of time. It is important to apply these changes before your website loses all visibility online.

  • It is time to stop focusing on the amount of traffic and start focusing on content consumption. As long as your content gets viewed, that is all that matters.


  • Embrace these changes and ensure your website is marked up in the right manner to enable its capability to handle instant articles.

The bottom line is that these changes are taking place because they are exactly what users want. By implementing these changes, you ensure that customers reach you and get the message you are sending them.

WordPress has a few great plugins such as Autoptimize and Better WordPress Minify that can be help you merge or even inline your CSS stylesheets.

  1. Mobile Devices Have Taken Over Desktops

In this day and age the number of mobile devices has continued to surpass the number of desktop computers by a wide margin. What this means is that most people are getting access to information through their smartphones or tablets.

As technology continues to advance, desktop computers are getting outdated. As a website owner, you should ensure your website is fully optimized for the web so that it is user friendly to mobile devices.

If you do not optimize your website you will continuously be losing potential customers who do not have the patience to zoom in to see what your website is about.

You should conduct a mobile friendly test to ensure your website appeals to mobile devices.

Take-home message:

Just by looking around you will discover that this is a fact. People all around carry their mobile devices everywhere and have their eyes constantly glued to them digging for information.

If you want your content to be discovered by internet users, you will have to keep up with the trend of ensuring your website is fully mobile optimized.

(Have you looked around lately? People of all ages live in their phones.)

There are 1.4 billion smartphone users and they are taking over the internet…
  1. The Hispanic Population is Shaping America

In the most recent census conducted in 2010 it was realized that the Hispanic population had the most substantial demographic impact on American population. There are now about 52 million Americans of Latin-American descent.

This number is projected to increase to 133 million by 2050 meaning that one in three Americans will be of Hispanic origin. The current Hispanic population already has a spending power of over $I trillion and is expected to increase in the coming years.

Take-home message:

The Hispanic market is no longer being viewed as a niche market, instead, it is now becoming a major market that marketers are starting to seriously focus on.

As a marketer or website owner it is up to you to ensure that your content can be understood by all customers despite the language they speak.

It is time to have your content translated to other languages such as Spanish that also have a significant market share.

  1. Online Users are Getting Lazy.

When searching for information on the internet users are determined to apply the least effort they can to get it. In response to this, Google has already come up with a tool that satisfies this ‘craving’.

Google Flight is a new tool created to help people with their travelling plans.

A study conducted by Google came to a realization that approximately 61% of travelers use search engines to carry out travelling plans. Google tapped on these results and came up with a way of channeling traffic from major travel sites such as Kayak and Orbitz into their new platform.

This has the impact of keeping users on Google’s search engines as well as feeding the appetite of lazy users, who now don’t have to put effort into visiting a new website to acquire travel information.

A new mobile interface (known as Destinations) has been created to further improve traveler’s user experience. This is what customers want. They can now take care of all their travel plans especially during vacation by using only their phones.

This is bad news for marketers since Google is determined to take over the travel industry searches by hijacking search traffic at all points.

Take-home message:

It is almost certain that Google will not stop at travel searches but will also extend to other industries. Google’s research never stops. If they find another lucrative industry, they will surely go for it and probably dominate it.

  • Focus on working within the structure of the platform. Google Destinations does not create content for their platform, they simply collect content from other websites. This implies there will still be a way to rank content.


  • Try out the platform. Since Google makes money through advertising, they will be working to monetize on this tool. If your business operates within this space you can buy ad space before the market gets crowded.
From a consumer point of view, this is pretty awesome. There’s no longer a need to search through slow websites – you can plan your entire vacation on your phone.
  1. Voice Searches are Taking Over Typed Searches
Image credit: Global Web Index

It has been found out that over the past 3 years, voice searches have been increasing by a rate of almost 50% each year. Another study has shown that 55% of teens and 41% of adults utilize voice search.

Voice searches source their information from the web and certain applications. However, voice queries work differently from typed queries.

  1. They apply natural language.
  1. They are not visual, only spoken out.

As the demand for voice searches grows, content will have to be structured in such a way that voice search queries are able to easily extract the information users need.

For marketers, this is an opportunity for them to start forcing it out through APIs such as echo and siri.

Take-home message:

Websites in the near future will start losing their value as more and more people switch to spoken queries, typed queries will become a thing of the past.

Kayak, a major player in the travel industry, has already started to put this information to good use by optimizing against voice search engines.

If your industry can benefit from implementing a similar tactic, as Kayak has done, then it’s time to go to work.

  1. The Era of Intelligent Machines is Here

Chatbots are one of the intelligent machines that Facebook is currently working on because they believe it is the future of social media, commerce and content uncovering.

Since most users are not willing to put any effort in searching for information, chatbots will take over this task and provide them with many intelligent options they can choose from at the touch of a button.

Take-home message:

Like many machines, chatbots are going to allow you run your business in an effective manner. A low cost bot will most likely be handling activities such as:

  1. Customer service
  1. Management of email
  1. Personalized searches
  1. Booking and purchasing

The impact that this will have on consumers is positive as bots will serve as their personal assistants.

As a marketer the best you can do is to simply embrace and readily adapt to this new trend when it finally takes over the current trend.

  1. VR (Virtual Reality) Headsets Will Push Content to a New Level

VR headsets are expected to change how gamers play games; however, the application of these headsets doesn’t stop there. They will also have a significant impact on commerce:

  1. One possible application is trying out items from stores without leaving your room
  1. You could also take virtual tours across the world in search of what you want
  1. You could also watch your favorite concert as if watching it from the front row

Take-home message:

Despite the possibility of all these applications, VR headsets are still a long way from actual mass application. Like the Google glass, it is still up to the market to make a choice whether they like the product or not.

  1. Content needs to be created in a smart way so that it will still work for you in future. This means your content should be easily consumable by users through instant messengers, VR headsets, phones and any future device.
  1. Marketing channels will keep evolving and we cannot afford to be picky. For example; Snapchat has a large influence on users and less competition compared to other platforms. It creates a great opportunity for marketers. Use such opportunities to your advantage it doesn’t matter whether you like the platform or not.

It all boils down to having a clear understanding of who your customers are and how they spend their time online. After knowing this, you can pick the best medium and create the right content for them.

These are just some of the changes that have been taking place in the web industry and acting on these changes will ensure that you stay ahead of your competitors.

If this article has been value to you, feel free to share it with your friends.



Quicker Indexing of Your Content Using the Fetch as Google Tool

Google Webmaster Tools have a variety of functions that provide a way of keeping your website running in a smooth manner. Two of the SEO tools that will prove to be of benefit to you when applied are the Site submission tool and the Crawl Errors tool.

The Fetch as Google option is found within the toolkit and has the option of giving users the chance to submit their URL to the index. Despite this unique and useful feature, a lot of SEO experts, bloggers and webmasters rarely use this tool. By utilizing this tool, you can get your new content discovered in the SERPs in a fast and convenient way.

Often when new web pages or blog posts are published, it usually can take weeks or even months for them to show up in the Google such results. Instead of just sitting back and waiting for them to show up a more skilled website owner will ensure that new content is included in their XML sitemap and then feed back their sitemap to Google.

This is where the Fetch as Google tool comes in. Normally, after applying this method Google is able to crawl the URL within a day. In some instances, webpages and SERPs can show up in less than 5 minutes of using the Fetch as Google tool.

This tool is very easy to use. It is worth it to try using it. Below are the steps to follow in order to use this tool.

Step 1: Visit Google Webmaster Tools

Before you are allowed to view the webmaster tools, you will be required to add a property by typing in your website’s URL into the box.

While in the Google Webmaster Tools home screen, select the domain name, click on the Crawl drop down menu, and then click on the Fetch as Google link.


Step 2: Fetch as Google

Type in your webpage or blog URL into the input field, excluding the domain name, and then click the FETCH button.

Step 3: Submit to Index

After clicking the fetch button, confirm that the Fetch status was successful. If you find that the Googlebot can successfully fetch your web page, you can proceed to submit that page to the Google index. To do this, click the Submit to index button.

There are two ways to go about the submission of indexes. It is possible to submit either the URL itself or the URL and all the pages that are linking from it. It is best to submit the URL if your page is new or has recently undergone some updates. However, if your website has undergone some major changes, it is in this case advisable to submit the URL and all linked pages. It is important to note that Google does not assure indexing of all URLs or pages linking to it.

Another advantage of using the Fetch as Tool is in that it allows Google comprehend that your website is the source of your content. Other websites can scrape your content and then outrank you if they extract your content from your RSS feed and get crawled by Googlebot sooner than your website.

Websites with frequent crawl rates or stronger domains are more likely to be considered as the sources of new content. By using the Fetch as Google tool, it is possible to get rid of scrapers who try to outrank you by hijacking your content.


From this article, we can conclude that the Fetch as Tool by Google is very important to bloggers, webmasters and marketers. It is a means not only to conveniently get content discovered faster on Google’s search engines but also a weapon that get rids of scrapers.

Every time we publish content our aim is to have our content appearing on search engine results within a short duration of time as we sit back and relax.

How to fix Image Upload Issue For Your E-commerce on WordPress-Step by Step for Beginners

When it comes to CMS (Content Management Systems), WordPress is the most commonly used platform in the world. Here are some of the reasons that make WordPress so convenient for most users:

  1. It has a huge community of users
  2. It is easy to use
  3. It is FREE!

From the image of the pie chart below we can see how WordPress compares to other content management systems in the world.

Even back in 2012, nine years after the release of WordPress, it was already running in approximately 60 million websites. It is one of the biggest things that have happened to the internet industry over the last decade.

However, WordPress just like many content management systems isn’t perfect. It has its limitations. It is important to note that WordPress was not designed by SEO experts. Hence, right from the start there were several SEO issues.

In this article we will focus on one widely used and commonly misunderstood WordPress feature, that is the WordPress Media Library. (Here second part 7 way to Image Web Optimization )

It is the set of forms and functions that WordPress uses to handle your images or other media that you upload to your website. If you do anything at all with WordPress or web images, it is important to learn about the media library because whenever you insert any sort of image into WordPress you will be interacting with the media library. The good news is that the media library is powerful and flexible, but it is also extremely confusing even to professionals, there is a lot of misinformation out there.

In this article we are going to learn exactly how to properly handle WordPress images both in terms of its usability and SEO perspective. After uploading images to WordPress, usually in the section illustrated by the following image, many WordPress users have a problem understanding what the information on right side of the Attachment Details section means.

A lot of people are familiar with a couple of fields but not many people understand them all. Therefore, what you will often find is that usually the website owner learns a little bit of SEO and then fills all the fields with some optimized text. After doing this, they think that they are in some way increasing the SEO liability of their web page.

Let us take a look at what all this information means, and exactly how to prepare SEO optimized images in WordPress. This is important because adding images to a post is probably the most commonly used WordPress feature.

To identify and fix the problems, we will use a before and after approach, using images. The ‘before section’ will show a users first attempt to fill the fields while the ‘after section’ will show how the problem can be fixed.

Problem 1: File name

In our example above our user actually started off with a common mistake which is the file name. The file name, 100_1623-1.jpg, is a standard digital camera numbering convention for naming a picture’s file name.

This means, that whenever you take a picture, your camera usually just gives the picture a numerical name like this one (100_1623-1.jpg). This is usually not helpful since these numbers are meaningless in terms of SEO.

Solution: To fix this issue, the website owner should have renamed the image to a name that has meaning for example; beige designer handbag.jpg.

Proper image SEO starts with optimizing the image file name first before you even upload the image file to your website.

Problem 2. Image File Size and Dimensions

Next, we look at the image file size and dimensions of the image.

The image has a file size of 4MB and dimensions of 4920 x 3264 pixels. This image file is huge, but it is what we expect from a quality digital camera. The problem is that it is not advisable to upload an image of this size to the internet because it causes a delay in the time taken to load a website. This means you could lose potential customers.

Solution: To fix this, the website owner should have used an image editor like PhotoShop to reduce the size of the image and optimized the image for the web

By using an image editor, the image can be down sized to 800 x 531 pixels and its size can also be reduced to 75KB which is about 1/140 the size of its previous size. This means that the image will load a lot faster for website users, and this is important because many studies have shown that you lose customers for every second your customers have to wait for your content to load.

Problem 3. Title and Description Fields

Let us now look at two other fields that need fixing, the title page and the description field.

These fields confuse a lot of people because when you go to insert this image into your post, whatever text you type here doesn’t actually show up on your post. That makes you wonder, what is the purpose of these fields?

The content from these two fields is used in a very specific location in WordPress called the WordPress attachment page. A lot of WordPress users are not familiar with this page because it is not usually very useful for the majority of websites. To get a look at this page you can click the bottom link titled ‘View attachment page

This link usually opens the URL of your WordPress attachment page. Below is an image of how it looks like in the WordPress default theme.

Scattered around the page is the title and description text as well as some other data. The question to ask yourself is, what is this page good for? The answer is… Nothing!

Many WordPress theme designers do not even bother to code this page. The reason is because in the vast majority of web business models, you typically never want a user to land on the attachment page. This is because the page has no navigation, no post content and most importantly there is no ‘BUY button’ on this page.

We want Google to be sending traffic to our actual product post page not our attachment page. We do not want potential customers to land on a page with no action steps on it. For most web business models the attachment page is useless at best, and at worst by utilizing it we are losing money. It is therefore considerable to set WordPress attachment page to no index.

Solution: With this in mind, what do we type in our title and description fields? The answer is… Absolutely Nothing! The website owner doesn’t want a user landing on the attachment page, so there is no reason to type anything into these fields.

Problem 4: Caption Field

The other field is the caption field.

The text in this field is typically rendered directly below the image on your post.

(Lager size image )

From a SEO perspective, this is important because the search engines will tend to associate the image with any text that is approximate to it, giving special emphasis to the nearest text and page title. From a site user’s perspective, readers have a tendency to read the text that is directly located below an image.

It is commonly stated that image captions are read 300% more than your article body content. You should be aware that what you write in the caption field has a very high probability of being read by your potential customers. Your web visitor might ignore everything else apart from your image caption.

Solution: Use this box to create an appealing and easily digestible descriptive sentence or two.

Problem 5: The Image Alt text Field

When it comes to the image Alt text field, whatever you type here will be used in the infamous, all attribute of the image html tag. Historically, the all attribute has been used to describe the image content to people with vision disability, people who cannot see the image for whatever reason. Since search engines are basically ‘blind’ as well, at least up until the quite recent A.I (Artificial Intelligence) advances, they use this attribute to get an idea of the image content too.

Solution: In this field, it is best to type in a keyword with a phrase that precisely describes the image using about five to ten words or so. As in the example illustrated by the image below, note that the text in this box, might be the kind of phrase that someone might type directly in into Google’s search box. We use a descriptive keyword optimized phrase to let both the search engine robots and the human users know what sort of product we are selling and where to find it on our website.

Discover How Google Ranks Websites

As a website owner it is important for you to understand how Google algorithms (robots) go about cataloging and ranking all the websites on the internet. It is important that you have some exposure to SEO methodology, because this article will be diving a bit deeper into the topic.

What is SEO?

SEO stands for Search Engine Optimization. It is the practice of optimizing your website so that search engines like Google are more likely to feature your web page above your competitor’s web page. People all around the world are searching for goods and services, and they use search engines as a vehicle to find the stuff that they want.

Google has a job of cataloging every webpage in the world. They try to guess what the searcher is looking for, based mainly on what the searcher types into the Google search box. The majority of people only click on one of the top 3 search results they see. Therefore, you want your webpage to be as high as possible in Google’s search engine results page, also called the SERP.

Why do we want to learn SEO?

It is important to learn SEO because the higher your website ranks on Google, the more traffic you get to your website. The more traffic you get to your website, the more money you make. (or simple hire us ^.^)

How do we do SEO?

SEO has become a pivotal component of the marketing strategy of millions of businesses. Back in the day, SEO was often synonymous to link building. Link building is the process of creating (or otherwise attaining) links on other people’s websites- that point to your website (also called Backlink building).

Backlinks are important because there are actually two types of users on the internet; humans and robots.

Google’s job is to catalog websites, and they do this by deploying thousands of robots (thousands of programs that search the internet and follow links all day and night). These robots have many names they are often referred to as; bots, spiders or crawlers.

Google’s program is often referred to as Googlebot. These programs ‘crawl’ the web, download websites and store the content at a Google data center. Eventually all this data is run through a complex algorithm that assigns various rankings and cataloging metrics to every URL on the internet.

Google database center image

How does Google assign a rank value to your website?

Traditionally, the website with the most links won. Every incoming link to a website is in Google’s ‘eyes’ seen as a vote for that website. The more a website accumulated more links the more it was able to rank higher and higher. Hence, the website would get more clicks and traffic, therefore making more money.

Google’s success was in part based on the observation that; the more links people made to a webpage, the more useful or trustworthy that webpage probably was. The math behind page rank is very complex. If you can get a lot of websites to link to your website, Google uses this metric as a hint that your website probably has a lot of value. This concept is described in the images below.

  1. If your webpage does not have other web-pages linking into it, then that is bad.

2. If your webpage has lots of webpages linking to it, then that is good.

  1. If the webpages linking to your webpage also has lots of webpages linking to it, then that is VERY good.

The top webpage is considered to be more important than all the other web pages that are below it.

It is therefore fairy easy to write a program (robot) to simply ‘crawl’ the internet and count the number of links pointing to each webpage. From here, an assumption that the webpage with the most accumulated links is probably better and should generally rank above websites that do not have many links is made. That was the original Google algorithm, the webpage with the most and strongest links wins.

However, there was a flaw to this algorithm. It could be and still can be ‘gamed’ by anyone willing to devote themselves to the monotonous chore of back-linking and creating thousands of links all pointing to the website they are trying to rank as illustrated in the image below.

It was this observation that inspired the birth of the billion dollar SEO industry, and a new generation of grey hat marketers who published an entire universe of hastily produced articles on the web that all contained links pointing back to their primary website. Many of these links contained cheaply written article content or were just pure spam links.

When Google realized this, they formed the web spam team, led by Matt Cutts, which started the indexing or otherwise punishing websites that were engaging in what Google termed as an artificial inflation of their Google’s search engine rankings.

Doing SEO used to entail a lot of backlinking, a lot of people claim that this is still what SEO entails. To some extent, this is true because the operation of Google’s algorithm still operates, on a large part, based on a webpage’s backlink value. Building this value is still a goal of off page SEO.

It is very hard these days to rank a webpage based purely on backlink construction. Google does have other factors that it uses to rank websites.

Three Factors (Other Than Backlinking) That May Affect Your Google Rankings

  1. Navigational Searches

These are people who use Google to find a particular website domain or brand. A big part of Google’s job is to return navigational search results. Often the world’s biggest search engine is not really a search engine at all, it is just a tool for looking up domain names.

In the early days if you didn’t type in the domain name of what you were looking for correctly, you could not get what you were looking for. An error message would usually appear as the one illustrated in the image below. You would usually have to start typing all over again.

These days people type in the brand name directly into the search box, the need to type in the full domain name is no longer necessary as in the figure below.

Modern search browsers interpret what a user types into the search box not as a domain but as a search query. The browser then sends this query to its default search engine, usually Google. Google receives this query and renders a search engine results page that usually does a good job of guessing which web destination you actually want to go to. The question to ask ourselves is how Google interprets queries from web browsers? Google places a lot of trust on the brand names that receive a lot of searches than those that receive very few searches.

In the example in the image below, the webpage,, would receive a higher ranking than the webpage,, because Google views as having better information since a lot of people interact with it.

This sort of data is very obvious and an easy to harvest trust metric for Google to utilize. It makes sense for Google to presume that should rank above

  1. Return Visitors

Google has access to the I.P addresses of unique users as they instigate navigational searches. This means that they can also log the I.P addresses of these unique visitors. Hence, Google knows which webpages have return visitors.

The picture above shows a highly simplified log file of four different users accessing a webpage via Google over a period of one week. We can see that the users visiting returned to the website again and again, multiple times throughout the week. However, those users visiting visited the website only once but never returned.

From these findings we can conclude that the website has better and useful information to users, meaning it is more trustworthy. Hence, Google is most likely to rank a website that receives more return visitors higher than its competitors. Return visitors is another metric that proves useful to Google when it comes to ranking websites.

  1. Pogo-Sticking

If a user clicks on a link of a webpage from the Google search results page, visits the webpage and does not like what he/she finds on that webpage, then returns to the Google search results page and clicks on another webpage link; this is referred to pogo-sticking.

Pogo-sticking is bad. Users clicking on a webpage link from Google’s search results then seconds later coming back to Google concludes that the users hate the website as it has no valuable information. For Google, this means this link should be moved down and the useful links should therefore be moved up.

Websites that satisfy the query of a Google user are ranked higher than the websites that don’t satisfy the query of Google users.


There are two distinctive pillars when it comes to SEO:

  1. Off-page SEO
  2. On-page SEO

This distinction needs to be stressed because the skills-sets employed to these two pillars are quite disjunctive. Backlink building, which is the practice of obtaining links from other peoples’ websites falls under the off-page SEO category. On-page SEO is different, it refers to the SEO work that you do “on-page” on your own website. On page SEO encompasses a few tasks that primarily include:

  1. Getting the Keywords on your webpage: your article content should contain keyword phrases that people actually type into Google when they are looking for your product or service.
  2. Create an easily navigable web architecture: Your content should be arranged in such a way that it is easily discoverable by a search engine such as Google.
  3. Plan a content strategy: apply a reliable editorial strategy to be employed when choosing which topics to write about on your website.

Essence of On-page SEO

The essence of On-page SEO is to strive to refer to your “product or service” using the same words that your customers use when they search for your “product or service”. For example if your customers are referring to a San Francisco florist, and the words ‘San Francisco’ and ‘florist’ do not appear on your webpage, then it is asking a lot for a Google algorithm, to know how to categorize and rank the website.

We, as humans, can see the picture of the golden gate bridge. We can also see a bunch of flowers for sale, so it is easy for us to infer that this is probably the website of a florist near San Francisco. However, in a machine’s ‘eyes’ this just looks like a page of colorful pictures and some text about flowers.

That is why we use Keyword tools to dig up lots of keywords that reveal what our potential customers, the human searchers on Google, are actually typing into Google. The method by which you insert these Keywords into Google is what comprises the art and science of on-page SEO.

On-page SEO is more preferable than Off-page SEO for several reasons:

  1. Change resistant: While Off-page SEO is a practice that is constantly in flux, on-page SEO is change resistant. Any changes made to the Google algorithms, such as the previous Panda and Penguin updates in most cases affect off-page SEO. However, such updates or changes do not affect on-page SEO.
  1. Control of Content: In on-page SEO you have full control of the contents of your website unlike for off-page SEO where all the content of the webpage is controlled by the owner of the website. Once you have published your backlink content on someone else’s server, you don’t retain any control of it. You are forever subject to the whims of the guy who owns the website.
  2. It is something you have to do anyway: whether you know SEO or not your website has to have some content on it. You have to put words on your website. Once you learn on-page SEO, the amount of time it takes to create an SEO optimized webpage is not really much more than the amount of time it takes to build a webpage that does not have any SEO optimization at all. Hence, if you have to build a webpage, it is good to have it built based on foundational SEO principles.

So, if you’ve got to build the site anyway, you might as well build it utilizing foundational SEO principles. These are the principles we discuss on this blog.. (if you keep learn it will become second-nature to you )

i .



iv . Or Simple hire us to do this task


7 Ways To Image Web Optimization


I want to share with you a pretty cool trick. I think this may be the most elusive technique in web development and web design in general. If I was asked to teach just one thing pertaining to web development, this would be the skill I would want each website owner to learn. Most website owners and developers probably have no idea about it.

But before we get to the trick, let us look at the various image file formats:

JPEG (Joint Photographic Experts Group)

This file format was developed by the Joint Photographic Experts Group. It works well for natural image types. Generally, JPEG is used when small image size is more important than image quality (such as web pages). JPEG is the most popular image format on the web.

GIF (Graphics Interchange Format)

GIF was developed by CompuServe in the computer 8-bit video era. This image format is ideal for logos, line arts and other simple images that have no gradients or varying color. GIF is good for web graphics that have limited color and transparent backgrounds.

PNG (Portable Network Graphics)

This file format was created fundamentally to replace and improve the GIF format. It is an image format specifically created for the web. PNG creates better quality images for the same kinds of limited-color images.


WebP is a new image format for the web developed by Google. It provides quality lossless and lossy compression for images on the web. These images are smaller, richer and make the web faster.

Now that you have an idea of the various image formats, let us get back to our trick.

So what is the trick?

The trick is based on understanding how nearly every website owner saves images on the web. How most website owners save images is WRONG!. So, how do I come to this conclusion?

This conclusion is based on the fact that when you visit the average website, your computer has to download various files from the website server. These files include; style-sheets, HTML, scripts, fonts, video, images and other files. The pie chart below represents the major file types (Image is courtesy of smashing magazine).

From the pie chart we can see that much of it is just text and computer code, but the biggest chunk of the download task is occupied by downloading of image files (almost two thirds of website files are image files). This is the area highlighted in red in the second pie chart below.

What this means is that whenever your readers or customers are waiting for your web page to finish loading, there’s a good chance they are waiting for your image files to finish downloading from your server to their computer.

What this means, is that big image files are a big problem. This becomes a huge problem because for every second your users have to wait, you are losing potential customers.

So, if you run a website, it’s in your best interest to learn the art and science of Image Web Optimization. Knowledge of image web optimization can cut your users wait time. Image web optimization focuses on reducing the size of the image, without compromising on the quality of the image.

If you learn how to do image web optimization correctly, you can reduce the image file size without any noticeable quality loss at all. It will be very difficult for the casual web visitor to perceive any difference in image quality. Web optimized images can be reduced to almost a fifth of the original image file.

Below is an image of a before and after effect of image web optimization on an image file.

As you can see from the image above the two images look identical but the image file sizes are not the same. The web optimized image which is 665 Kilobytes (after), is almost a fifth of the original image 4,583 Kilobytes (before). Also notice how the web optimized image has not lost its quality.

Here is another example.

The original image is 2,136 Kilobytes and the web optimized image is 623 Kilobytes. The original image has been cut down to almost a third.

I know by now you are wondering… What is the trick? How do I do it?


We will use the above Google+ default image in our example. Remember that the original Google+ image is 2,136 Kilobytes.

To optimize this image, we are going to use Adobe Photoshop Elements as our image editing software since it is the industry’s standard.

Step 1: Open the adobe Photoshop software on your computer and then open the image you want to web optimize (In our case the 2,136 Kilobyte Google+ JPEG image file).

Side note :

Step 2: Click on the file menu then scroll down to “Save for Web…” or the short cut keys combination; (Alt+Shift+Ctrl+S )

This opens up the Photoshop Save for Web interface below.

From this interface, we can see our original image at the top and the web optimized image we are about to create at the bottom.

On the top right corner of our interface, you will notice that we have our image format set to JPEG, which is appropriate for this image. On the image format drop down menu, you can select a different image format. Proper Image format selection is a huge topic in itself.

Step 3: Just below the image format selector, move the image quality slider.

By moving the slider all the way to zero quality, the size of the image will become small but the image will not look good. It will look a bit blurry, which is not what we want.

By moving the quality slider all the way to 100 the image will look good, but the image file will be very large. This is also not what we want.

In image web optimization, our goal is to keep the image size small without making our image quality “suffer”. We have to keep moving the image quality down until we reach to a point where we are comfortable with both the size and the quality of our image.

 Step 4: Move the image quality slider to a point where image size is reduced without affecting image quality.

In our example above, the quality of our image begins to “suffer” at below 50%. So for example, I can set the image quality at around 65%. At 65%, the image size is small and the quality of the image is not affected.

The bottom left corner tells us what the image size will be when we save it. In our example, the final image size is 572.9 Kilobytes. This is about a quarter the size of the original image.

Step 5: Save the image.

At the bottom of the Photoshop Save for Web interface, you will find a save button. Click on this button to save your web optimized image.

From this tutorial, we can appreciate the value of image web optimization. If we assume every image on the web page had similar poor optimization, then we can cut the file load of that web page in half using this technique.

I hope from this article you have learnt more on web optimization. I encourage you to apply this technique to your own website, it will make a BIG difference for your website users.

Improving web performance and giving a better experience to our users is our job as developers and designers.




Cracking the SEO Code: search optimization tips

Search optimization tips

 So what has changed in SEO in the last 5 years? We actually had a very substantial shift in the SEO world. The ranking algorithms have become tremendously more complex as now we have more inputs and more sophistication.

There are developments in the use of data signals. For example, if 500 friends were to get on their phone and search for a keyword phrase and click on result number 9, that result will move up on Google pretty quickly. That result will fall back down after a little while but Google is taking into account some user signals that previously they didn’t seem to notice.

They’re also looking at pogo sticking: they perform a research, get to your startup website and then they go back to the search results. That’s a bad signal telling Google that this person was unsatisfied with the query and the answer that you gave.

pogo sticking
Google also has a supernatural skill to spot editorial versus manipulative links and this advancement has been drastic over the last four years. Today is a hundred times harder to violate Google’s rules when it comes to acquiring links. Some sophisticated black hats still manage to do it temporarily but if you’re building your brand, it’s not worth sacrificing your site.

Keyword matching has become query intent matching. Keyword matching has become tremendously more complex. Google has things like entities they look at, and far far better topic modelling algorithms.

There is also a ton of integration according to device, location, history, and Google+. If you own gmail account, a Google+ account is automatically created for you, and your search results will be influenced if you are signed in into your gmail account.

SEO has sort of shifted as well. Over the last few years SEO has gone from a job title to a part of a job description. Web designers, marketing executives, marketing managers, marketing strategists, and website managers include SEO in their job description. The only place where SEO is still very popular in the job title is in India.

SEOs found out that doing their jobs is not enough and they have to broaden out. What’s very frustrating today is that this non-editorial links (web spam) are more likely to cause problems because Google has launched a spam monitoring system inside webmaster tools , and it’s now all of our responsibilities to go and disavow all the spammy links that are pointing to our site.

If those links are not disavowed, bad things can happen to our site. Google likes to say that it’s incredibly hard for others to do negative SEO to our site without our knowledge. Unfortunately, there is still quite a big of risk of that happening. Because of this, SEOs are forced to manage, maintain, and keep on monitoring their sites all the time.

Google is also shortening the searcher’s path and dis-intermediating content creators. This is affecting SEOs in dramatic ways. For some search results, Google is taking over by listing their answers at the top of the first page; Google didn’t even wrote them as those answers were scraped and surfaced on the page through their UI.

The only good thing about it is Google using it to get people more addicted to search and thus make them search more often. The average number of queries per searches has dramatically gone up over the last five years, and the number of new searchers has gone up dramatically as well.

By the end of this year, Google expects more searches on mobile than on the desktop. Even if Google cannibalizes 25%, 30% of all the searches that will happen, there will still be more opportunity in search, in SEO than ever has been. SEO maintains that 80/20 split with paid ads so SEO still has a huge share there.

The social media pervasiveness has enlarged the field of influencers. Today, In US 72% of regular web users can influence and amplify content.

Also, disappearing data make SEO harder measure and improve. A lot of the search traffic today is being improperly attributed as direct traffic. These changes are having a relatively substantial impact but the SEO going forward will be more about:

– getting design and user experience right so that people don’t bounce when they land on our page because that can affect our rankings;
– getting our accessibility right because Google announced that https will now give you a slight benefit in the rankings if you possess a security certificate;
– press and PR, web spam, and social media have a direct influence on SEO;

SEO is hard. Our job as marketers is to connect these things up, to marry our strengths, the tactics that will be most effective for us with the opportunities that exist in search today. SEO is not dead. Google sends ten times more amount of traffic than Facebook every day. No tactics will work for everyone but we should be aware of all of them.

Going forward, keyword research should be done by looking at several data munched up. Almost everyone is still using Adwords for keyword research, meaning if you go outside of that and you find keyword opportunities that are not in Adwords, chances are the competition will be less competitive. Many search querries with hundreds or thousands monthly searches don’t appear in Adwords. This super simplistic and completely free tool at helps making Google suggest based research way easier and is one of the best research tools in the SEO world.

I strongly recommend for you to use your customer data. Sometimes your customers are not yet on your website. For early stage startups you don’t have customers talking in your help and support forums and communities about the language that they use to describe your products, services, and the problems they have; but they exist somewhere on the web, on sites like Reddit for example.

When it comes to content creation, your content should be:

1. Unique – it’s notspreaded on the web in a search-indexable format.
2. Relevant – actually contain terms and phrases that Google will translate as on topic.
3. Helpful – should resolve the searcher’s queries efficiently.
4. Uniquely Valuable – provides information that is not available anywhere else on the web.
5. Great User Experience – easy to browse on any device, fast to load, and pleasurable.

(Two of my favourite UX resources at the moment, to get non-UX people thinking about UX:

  1. Designing User Interfaces for Your Mother –
  2. The user is drunk – )

Lots of startup websites often create content intended to target links and shares from a community that’s not relevant to their customers, or what they actually do. And some of that is ok but if you’re doing consistent off-topic click bait, or link bait, Google will punish your site for that. And when they’re doing it manually, penalties will last a long time. Google also hates guest posting and commenting as well.

The truth is that Google doesn’t take in consideration links that you can control and build. They only want to count the links that you editorially “earn” and that’s very frustrating, and a lot harder.

Today, social media is primarily impacting SEO indirectly but social shares will expose content to people who might amplify and link to it in good useful ways. The exception is Google+. If you were to get all your potential customers to follow you on Google+, you would rank number one for every search they perform. All you have to do is share your content on Google+. It’s cheating for SEO, and Google allows that cheat because it helps them acquire more customers for Google+. This can last for months.

The fundamental problem with content marketing is that it only works if you are close to the best in the world at it, and that is frustrating because not many people are but many think that they are. If you want to invest in content, here are some rules you should follow:

A) Strategic and Relevant – it must follow you goals and elevate your brand.
B) Targets Likely Amplifiers – if it’s unlikely that people will spread your content, then don’t bother publishing it.
C) Your content must be better than anything your competition has to offer in order to achieve success.

You can utilize Buzzsumo to show you all the posts on a certain topic across the entire web over the last year or so that have earned the most social shares across all networks. From this data you can really get a sense of which content was the most appreciated and useful.

You should not measure your ROI just by giving credit to the last channel that sent you the click that converted. Customers that convert the best usually come to a site multiple times from various sources (social media sites, direct visits, email, etc.) before they actually convert. If you gave the last channel all the credit, you would never invest in these assisting channels like social, like email, like content that provide huge benefit to you.

Today you need to use evaluate the performance of your traffic by page (or section) instead of traffic and performance by keyword. Tools such as Conductor, or Moz Analytics can help you predict which keywords send the most traffic to a certain page.

SEO is more difficult these days, but those who work harder and smarter have a greater chance to succeed.

here is a cover of all the episode shark tank

Huge audience wave ahead dive in on shark tank episodes

here is a picture of all the shark tank members sitting
shark tank episodes

It takes a lot of planning and insight for it to work to make ABC’s Shark tank site a success.

Such as success is not something that can be mathematically calculated and expected to work out without putting in place any safeguards.

This article looks at the planning that made it possible to successfully manage the large website traffic handled by

ABC’s Shark tank Website

. This same steps can be replicated to ensure that your website do not crash due there is a cover of all the episode shark tank o lack of planning.

How to prepare for a traffic surgePreparing for a surge in website traffic is like preparing for a basketball game. The surge in website traffic is your opponent in this analogy.
To successfully prepare for ABC’s Shark tank site’s surge in traffic, they did the following:
• They assessed their own strengths and weaknesses as well as that of their opponent.
• They picked their plays and players earlier on before the play.
• They understand what visitors they will be dealing with, what devices they will be using, how long they will use them on the site, and what they will be doing.

Once the game is on, it is about playing systematically. In this case, all website support must be handled systematically. It is like studying the playbook and learning the secrets behind success. Then during the game, you apply them. Once the traffic surge has subsided, it is time to look at what went well, and what went wrong. This is the same thing that happens after a game, and what is practiced at ABC’s Shark tank online.
Talk to experienced peopleExperience is always welcome when facing something that is not so certain. Whether you are an app developer, or a game developer, get as much detail as you can from those who have handled traffic surges before. The team behind ABC’s Shark tank website is well experienced and they apply their experience accordingly.
Assemble an excellent teamABC’s Shark tank site engaged a hosting company that has a fulltime engineer, a developer, DevOps, a business decision maker, and a project manager. They used people who have had experience with high traffic events in all these positions if you can. Otherwise at the minimum you should have the hosting company’s engineer and the DevOps team.

A developer who will follow all instruction given will work just fine. Even without any experience. However, a problem in the code can be devastating. At ABC’s Shark tank site, they use excellent developer all through.
Common mistakesMany executives tend to lump IT together. There is a difference between server configurations, programming (both web and backend), network configuration, and databases. It is common in small companies for one person to manage it all, but to successfully deal with web (or shark tank products) surge traffic, each skill set should be clearly defined and assigned to members of your team. At ABC’s Shark tank site, all roles are well defined and allocated to each team member.

Hiring tipsAt ABC’s Shark tank site, all those who were hired are competent and skilled professionals. When you are interviewing for a position and you get a lot of caveats for an answer. You should ask what it would take to get an answer without the caveats.

Example:Are you confident you can configure the servers?
Right answer:Yes, I understand all the technologies involved and I can handle it.
No really. I understand what needs to be done but am not an expert; we may need a second set of eyes.
Wrong answer:Yes, but am not the one who configured the server.
HostingABC’s Shark tank online was successful because it did not use shared hosting or a single cloud server. It also did not use cloud managed service either. The challenge of a cloud managed service is that you can only alert them but you have no control over the code or server configuration. A virtual server is also not ideal as you need several servers, not a fraction of a server. Virtual Private Servers (VPN), also pose the danger of sudden crash caused by other sites that share them. ABC’s Shark tank site uses multiple servers.
Website or software testingYour website during a traffic surge is like a racing car at high speed. You do not need antennas or something that could cause drag. After all you will have taken out the radio to keep the car light. However, when driving at low speeds, a radio and a giant antenna won’t make a significant impact, if any. ABC’s Shark tank site has only content that is relevant to its business objective.

Its pages were tested to ensure they load the videos under four seconds.
ABC’s Shark tank site’s database and server were tested to make sure they were ready and their resilience established to ensure it can support a large load for a prolonged period of time.
The ABC’s Shark tank site capitalized on stuff that can close the deal only. ABC’s Shark tank site has no items that won’t move people to check out, at least for the traffic surge duration. ABC’s Shark tank site does not waste valuable space of long introductory videos.
ABC’s Shark tank site was load tested for various events such as logins, sales, and page loads. An emergency plan was laid out, to take care of the unexpected.
They have a HTML landing page with optimized images and key marketing objectives. They use Content distribution Networks (CDN). This ensures that the server is not burdened.

A notice for the CEO The ABC’s Shark tank site executives take technical advice into consideration in place in making business decisions. They understand that, traffic surges are not a continuous thing but more of a moment event. They understand all challenges that might come and are supportive of everything the technical team does to make sure that all the show is a success each time. They also take responsibility of any decisions they make and do not blame the technical team for any hitches, if any. Overall, ABC’s Shark tank site’s success in handling such high traffic surges mainly due to the planning techniques discussed above rather than use of high end equipment alone.alone.