Welcome to www.minisinbox.com

Web Hosting - FTP and Other File Transfer Tools Anything related to the Internet or computers is bound to introduce technical issues pretty soon. One of the earliest that novice web site owners encounter is FTP, which is an acronym for File Transfer Protocol. Seeing it spelled out, it's easy to see why those in the know quickly move to speaking in short hand. The reason web site owners soon will (or need to) become familiar with FTP is obvious to anyone who has built a site on a remote server. You have to have some way of getting the files to the remote computer and FTP is one of the most common tools. It's also one of the simplest and most efficient. FTP is composed of two parts: the client software and the server software. It's similar, in a way, to talking to someone on the phone who writes down everything you say. You (the client) make a request ('transfer this file to the server') and the listener (the server) takes the request and acts on it. That request to copy a file from a local computer to the remote one is carried out (often 'under the covers') by a PUT command, as in PUT this there. You create the web page (in the form of a file) and then PUT the file on the server. To move a file in the opposite direction, from the remote server to your local computer, your client software issues a GET command. Many FTP clients have graphical interfaces, similar to Windows Explorer, that allow you to drag-and-drop or otherwise copy the file without ever seeing the actual commands that carry it out. But it's helpful sometimes to know what goes on underneath. In tricky cases it can be an advantage to use a command line interface (in Windows, the 'DOS box', with a similar interface familiar to most Linux users). Knowing the commands and being able to use them in the command line form can sometimes help you diagnose what is going on when the graphical tools misbehave. But FTP is not the only way to get a file from here to there. In fact, your browser moves files around from a remote computer to your local one all the time. In most cases, when you type in or click on a URL, what happens under the covers is in essence a file transfer process. The web page is transferred from the web server to your local computer then displayed by the browser. Alternatively, you can sometimes even email a web page/file from your local computer to the remote server, then use an email client on the server itself to get the file and put it in a folder. That requires that you have some form of access to the remote computer. But there are many ways of doing that, such as in-built utilities in the operating system or using commercial remote control programs. Those alternatives can be helpful to know in cases where the FTP file transfer process is misbehaving. Having more than one way to accomplish the task helps you diagnose what might be going wrong. It also helps you get the job done when the usual tools aren't cooperating. The more you learn about these sometimes puzzling acronyms, the easier you can accomplish your own goals.

Web Hosting - Sharing A Server – Things To Think About You can often get a substantial discount off web hosting fees by sharing a server with other sites. Or, you may have multiple sites of your own on the same system. But, just as sharing a house can have benefits and drawbacks, so too with a server. The first consideration is availability. Shared servers get re-booted more often than stand alone systems. That can happen for multiple reasons. Another site's software may produce a problem or make a change that requires a re-boot. While that's less common on Unix-based systems than on Windows, it still happens. Be prepared for more scheduled and unplanned outages when you share a server. Load is the next, and more obvious, issue. A single pickup truck can only haul so much weight. If the truck is already half-loaded with someone else's rocks, it will not haul yours as easily. Most websites are fairly static. A reader hits a page, then spends some time skimming it before loading another. During that time, the server has capacity to satisfy other requests without affecting you. All the shared resources - CPU, memory, disks, network and other components - can easily handle multiple users (up to a point). But all servers have inherent capacity limitations. The component that processes software instructions (the CPU) can only do so much. Most large servers will have more than one (some as many as 16), but there are still limits to what they can do. The more requests they receive, the busier they are. At a certain point, your software request (such as accessing a website page) has to wait a bit. Memory on a server functions in a similar way. It's a shared resource on the server and there is only so much of it. As it gets used up, the system lets one process use some, then another, in turn. But sharing that resource causes delays. The more requests there are, the longer the delays. You may experience that as waiting for a page to appear in the browser or a file to download. Bottlenecks can appear in other places outside, but connected to, the server itself. Network components get shared among multiple users along with everything else. And, as with those others, the more requests there are (and the longer they tie them up) the longer the delays you notice. The only way to get an objective look at whether a server and the connected network have enough capacity is to measure and test. All systems are capable of reporting how much of what is being used. Most can compile that information into some form of statistical report. Reviewing that data allows for a rational assessment of how much capacity is being used and how much is still available. It also allows a knowledgeable person to make projections of how much more sharing is possible with what level of impact. Request that information and, if necessary, get help in interpreting it. Then you can make a cost-benefit decision based on fact.

Web Hosting - Sharing A Server – Things To Think About You can often get a substantial discount off web hosting fees by sharing a server with other sites. Or, you may have multiple sites of your own on the same system. But, just as sharing a house can have benefits and drawbacks, so too with a server. The first consideration is availability. Shared servers get re-booted more often than stand alone systems. That can happen for multiple reasons. Another site's software may produce a problem or make a change that requires a re-boot. While that's less common on Unix-based systems than on Windows, it still happens. Be prepared for more scheduled and unplanned outages when you share a server. Load is the next, and more obvious, issue. A single pickup truck can only haul so much weight. If the truck is already half-loaded with someone else's rocks, it will not haul yours as easily. Most websites are fairly static. A reader hits a page, then spends some time skimming it before loading another. During that time, the server has capacity to satisfy other requests without affecting you. All the shared resources - CPU, memory, disks, network and other components - can easily handle multiple users (up to a point). But all servers have inherent capacity limitations. The component that processes software instructions (the CPU) can only do so much. Most large servers will have more than one (some as many as 16), but there are still limits to what they can do. The more requests they receive, the busier they are. At a certain point, your software request (such as accessing a website page) has to wait a bit. Memory on a server functions in a similar way. It's a shared resource on the server and there is only so much of it. As it gets used up, the system lets one process use some, then another, in turn. But sharing that resource causes delays. The more requests there are, the longer the delays. You may experience that as waiting for a page to appear in the browser or a file to download. Bottlenecks can appear in other places outside, but connected to, the server itself. Network components get shared among multiple users along with everything else. And, as with those others, the more requests there are (and the longer they tie them up) the longer the delays you notice. The only way to get an objective look at whether a server and the connected network have enough capacity is to measure and test. All systems are capable of reporting how much of what is being used. Most can compile that information into some form of statistical report. Reviewing that data allows for a rational assessment of how much capacity is being used and how much is still available. It also allows a knowledgeable person to make projections of how much more sharing is possible with what level of impact. Request that information and, if necessary, get help in interpreting it. Then you can make a cost-benefit decision based on fact.