Device Servers Tutorial
Device Servers Tutorial
Device Servers Tutorial
For easy reference, please consult the glossary of terms at the end of this paper.*
The ability to manage virtually any electronic device over a network or the Internet is
changing our world. Companies want to remotely manage, monitor, diagnose and control
their equipment because doing so adds an unprecedented level of intelligence and efficiency
to their businesses.
With this trend, and as we rely on applications like e-mail and database management for core
business operations, the need for more fully-integrated devices and systems to monitor and
manage the vast amount of data and information becomes increasingly more important. And,
in a world where data and information is expected to be instantaneous, the ability to manage,
monitor and even repair equipment from a distance is extremely valuable to organizations in
every sector.
This need is further emphasized as companies with legacy non-networked equipment struggle
to compete with organizations equipped with advanced networking capabilities such as
machine-to-machine (M2M) communications. Theres no denying that advanced
networking provides an edge to improving overall efficiencies.
This tutorial will provide an overview and give examples of how device servers make it easy
to put just about any piece of electronic equipment on an Ethernet network. It will highlight
the use of external device servers and their ability to provide serial connectivity for a variety
of applications. It will touch on how device networking makes M2M communication possible
and wireless technology even more advanced. Finally, as any examination of networking
technologies requires consideration of data security, this paper will provide an overview of
some the latest encryption technologies available for connecting devices securely to the
network.
From the evolutionary standpoint, the networking infrastructure of many sites has only
recently been developed to the point that consistent and perceived stability has been obtained
as users and management have become comfortable with the performance of the network,
they now focus on how they can maximize corporate productivity in non-IS capacities.
Device server technology solves this problem by providing an easy and economical way to
connect the serial device to the network.
Lets use the Lantronix UDS100 Device Server as an example of how to network a RAID
controller serial port. The user simply cables the UDS100 s serial port to the RAID
controllers serial port and attaches the UDS100s Ethernet interface to the network. Once it
has been configured, the UDS100 makes that serial port a networked port, with its own IP
address. The user can now connect to the UDS100 s serial port over a network, from a PC or
terminal emulation device and perform the same commands as if he was using a PC directly
attached to the RAID controller. Having now become network enabled, the RAID can be
managed or controlled from anywhere on the network or via the Internet.
The key to network-enabling serial equipment is in a device servers ability to handle two
separate areas:
1. the connection between the serial device and the device server
2. the connection between the device server and the network (including other network
devices)
Traditional terminal, print and serial servers were developed specifically for connecting
terminals, printers and modems to the network and making those devices available as
networked devices. Now, more modern demands require other devices be network-enabled,
and therefore device servers have become more adaptable in their handling of attached
devices. Additionally, they have become even more powerful and flexible in the manner in
which they provide network connectivity.
Device servers allow independence from proprietary protocols and the ability to meet a
number of different functions. The RAID controller application discussed above is just one of
many applications where device servers can be used to put any device or machine on the
network.
PCs have been used to network serial devices with some success. This, however, required the
product with the serial port to have software able to run on the PC, and then have that
application software allow the PCs networking software to access the application. This task
equaled the problems of putting Ethernet on the serial device itself so it wasnt a satisfactory
solution.
To be successful, a device server must provide a simple solution for networking a device and
allow access to that device as if it were locally available through its serial port. Additionally,
the device server should provide for the multitude of connection possibilities that a device
may require on both the serial and network sides of a connection. Should the device be
connected all the time to a specific host or PC? Are there multiple hosts or network devices
that may want or need to connect to the newly-networked serial device? Are there specific
requirements for an application which requires the serial device to reject a connection from
the network under certain circumstances? The bottom line is a server must have both the
flexibility to service a multitude of application requirements and be able to meet all the
demands of those applications.
Lantronix has built on its long history and vast experience as a terminal, print and serial
server technology company to develop more functionality in its servers that cross the
boundary of what many would call traditional terminal or print services. Our technology
provides:
Ease of Use
As an independent device on the network, device servers are surprisingly easy to manage.
Lantronix has spent years perfecting Ethernet protocol software and its engineers have
provided a wide range of management tools for this device server technology. Serial ports are
ideal vehicles for device management purposes a simple command set allows easy
configuration. The same command set that can be exercised on the serial port can be used
when connecting via Telnet to a Lantronix device server.
An important feature to remember about the Lantronix Telnet management interface is that it
can actually be run as a second connection while data is being transferred through the server
this feature allows the user to actually monitor the data traffic on even a single-port servers
serial port connection while active. Lantronix device servers also support SNMP, the
recognized standard for IP management that is used by many large network for management
purposes.
Finally, Lantronix has its own management software utilities which utilize a graphical user
interface providing an easy way to manage Lantronix device servers. In addition, the servers
all have Flash ROMs which can be reloaded in the field with the latest firmware.
Data Acquisition
M2M
Wireless Communication/Networking
Factory/Industrial Automation
Security Systems
Bar Code Readers and Point-of-sale Scanners
Medical Applications
Data Acquisition
Microprocessors have made their way into almost all aspects of human life, from automobiles
to hockey pucks. With so much data available, organizations are challenged to effectively and
efficiently gather and process the information. There are a wide variety of interfaces to
support communication with devices. RS-485 is designed to allow for multiple devices to be
linked by a multidrop network of RS-485 serial devices. This standard also had the benefit of
greater distance than offered by the RS-232/RS-423 and RS-422 standards.
However, because of the factors previously outlined, these types of devices can further
benefit from being put on an Ethernet network. First, Ethernet networks have a greater range
than serial technologies. Second, Ethernet protocols actually monitor packet traffic and will
indicate when packets are being lost compared to serial technologies which do not guarantee
data integrity.
Lantronix full family of device server products provides the comprehensive support required
for network enabling different serial interfaces. Lantronix provides many device servers
which support RS-485 and allow for easy integration of these types of devices into the
network umbrella. For RS-232 or RS-423 serial devices, they can be used to connect
equipment to the network over either Ethernet or Fast Ethernet.
Two extremely important and useful technologies for communication that depend heavily on
device servers are M2M and wireless networking.
Maximized efficiency
More streamlined operations
Improved service
Lantronix Device Servers enable M2M communications either between the computer and
serial device, or from one serial device to another over the Internet or Ethernet network using
serial tunneling. Using this serial to Ethernet method, the tunnel can extend across a
facility or to other facilities all over the globe.
M2M technology opens a new world of business intelligence and opportunity for
organizations in virtually every market sector. Made possible through device servers, M2M
offers solutions for equipment manufacturers, for example, who need to control service costs.
Network enabled equipment can be monitored at all times for predictive maintenance. Often
when something is wrong, a simple setting or switch adjustment is all that is required. When
an irregularity is noted, the system can essentially diagnose the problem and send the
corrective instructions. This negates a time-consuming and potentially expensive service call
for a trivial issue. If servicing is required, the technician leaves knowing exactly what is
wrong and with the proper equipment and parts to correct the problem. Profitability is
maximized through better operating efficiencies, minimized cost overruns and fewer wasted
resources.
M2M technology also greatly benefits any organization that cannot afford downtime, such as
energy management facilities where power failures can be catastrophic, or hospitals who
cant afford interruptions with lives at stake. By proactively monitoring networked-enabled
equipment to ensure it is functioning properly at all times, business can ensure uptime on
critical systems, improve customer service and increase profitability.
Wireless Networking
Wireless networking, allows devices to communicate over the airwaves and without wires by
using standard networking protocols. There are currently a variety of competing standards
available for achieving the benefits of a wireless network. Here is a brief description of each:
For more in-depth information, please consult the Lantronix wireless whitepaper which is
available online.
Wireless device networking has benefits for all types of organizations. For example, in the
medical field, where reduced staffing, facility closures and cost containment pressures are
just a few of the daily concerns, device networking can assist with process automation and
data security. Routine activities such as collection and dissemination of data, remote patient
monitoring, asset tracking and reducing service costs can be managed quickly and safely with
the use of wireless networked devices. In this environment, Lantronix device servers can
network and manage patient monitoring devices, mobile EKG units, glucose analyzers, blood
analyzers, infusion pumps, ventilators and virtually any other diagnostic tool with serial
capability over the Internet.
Forklift accidents in large warehouses cause millions of dollars in damaged product, health
claims, lost work and equipment repairs each year. To minimize the lost revenue and increase
their profit margin and administrative overhead, a company has utilized wireless
networking technology to solve the problem. Using Lantronix serial-to-802.11 wireless
device server the company wirelessly network-enables a card reader which is tied to the
ignition system of all the forklifts in the warehouse. Each warehouse employee has an
identification card. The forklift operator swipes his ID card before trying to start the forklift.
The information from his card is sent back via wireless network to computer database and it
checks to see if he has proper operators license, and that the license is current. If so, forklift
can start. If not the starter is disabled.
For shops that are running automated assembly and manufacturing equipment, time is money.
For every minute a machine is idle, productivity drops and the cost of ownership soars. Many
automated factory floor machines have dedicated PCs to control them. In some cases,
handheld PCs are used to reprogram equipment for different functions such as changing
computer numerically controlled (CNC) programs or changing specifications on a bottling or
packaging machine to comply with the needs of other products. These previously isolated
pieces of industrial equipment could be networked to allow them to be controlled and
reprogrammed over the network, saving time and increasing shop efficiency. For example,
from a central location (or actually from anywhere in the world for that matter) with network
connectivity, the machines can be accessed and monitored over the network. When
necessary, new programs can be downloaded to the machine and software/firmware updates
can be installed remotely.
One item of interest is how that input programming is formatted. Since many industrial and
factory automation devices are legacy or proprietary, any number of different data protocols
could be used. Device servers provide the ability to utilize the serial ports on the equipment
for virtually any kind of data transaction.
Lantronix accounts for this need by supporting RTS/CTS flow control on its DB25 and RJ45
ports. Lantronix device servers handle everything from a simple ASCII command file to a
complex binary program that needs to be transmitted to a device.
Security Systems
One area that every organization is concerned about is security. Card readers for access
control are commonplace, and these devices are ideally suited to benefit from being
connected to the network with device server technology. When networked, the cards can be
checked against a centralized database on the system and there are records of all access
within the organization. Newer technology includes badges that can be scanned from a
distance of up to several feet and biometric scanning devices that can identify an individual
by a thumbprint or handprint. Device servers enable these types of devices to be placed
throughout an organizations network and allow them to be effectively managed by a
minimum staff at a central location. They allow the computer controlling the access control to
be located a great distance away from the actual door control mechanism.
One more area of security systems that has made great strides is in the area of security
cameras. In some cases, local municipalities are now requesting that they get visual proof of a
security breach before they will send authorities. Device server technology provides the user
with a host of options for how such data can be handled. One option is to have an open data
pipe on a security camera this allows all data to be viewed as it comes across from the
camera. The device server can be configured so that immediately upon power-up the serial
port attached to the camera will be connected to a dedicated host system.
Another option is to have the camera transmit only when it has data to send. By configuring
the device server to automatically connect to a particular site when a character first hits the
buffer, data will be transmitted only when it is available.
One last option is available when using the IP protocol a device server can be configured to
transmit data from one serial device to multiple IP addresses for various recording or archival
concerns. Lantronix device server technology gives the user many options for tuning the
device to meet the specific needs of their application.
Scanning Devices
Device server technology can be effectively applied to scanning devices such as bar code
readers or point-of-sale debit card scanners. When a bar code reader is located in a remote
corner of the warehouse at a receiving dock, a single-port server can link the reader to the
network and provide up-to-the-minute inventory information. A debit card scanner system
can be set up at any educational, commercial or industrial site with automatic debiting per
employee for activities, meals and purchases. A popular amusement park in the United States
utilizes such a system to deter theft or reselling of partially-used admission tickets.
Medical Applications
The medical field is an area where device server technology can provide great flexibility and
convenience. Many medical organizations now run comprehensive applications developed
specifically for their particular area of expertise. For instance, a group specializing in
orthopedics may have x-ray and lab facilities onsite to save time and customer effort in
obtaining test results. Connecting all the input terminals, lab devices, x-ray machines and
developing equipment together allows for efficient and effective service. Many of these more
technical devices previously relied upon serial communication or worse yet, processing being
done locally on a PC. Utilizing device server technology they can all be linked together into
one seamless application. And an Internet connection enables physicians the added advantage
of access to immediate information relevant to patient diagnosis and treatment.
Larger medical labs, where there are hundreds of different devices available for providing test
data, can improve efficiency and lower equipment costs by using device server technology to
replace dedicated PCs at each device. Device servers only cost a fraction of PCs. And, the
cost calculation is not just the hardware alone, but the man-hours required to create software
that would allow a PC-serial-port-based applications program to be converted into a program
linking that information to the PCs network port. Device server technology resolves this
issue by allowing the original applications software to be run on a networked PC and then use
port redirector software to connect up to that device via the network. This enables the medical
facility to transition from a PC at each device and software development required to network
that data, to using only a couple of networked PCs doing the processing for all of the devices.
There are two main types of encryption: asymmetric encryption (also known as public-key
encryption) and symmetric encryption. There are many algorithms for encrypting data based
on these types.
AES
AES (Advanced Encryption Standards) is a popular and powerful encryption standard that
has not been broken. Select Lantronix device servers feature a NIST-certified implementation
of AES as specified by the Federal Information Processing Specification (FIPS-197). This
standard specifiesRijndael as a FIPS-approved symmetric encryption algorithm that may be
used to protect sensitive information. A common consideration for device networking
devices is that they support AES and are validated against the standard to demonstrate that
they properly implement the algorithm. It is important that a validation certificate is issued to
the products vendor which states that the implementation has been tested. Lantronix offers
several AES certified devices including the AES Certified SecureBox SDS1100 and the AES
Certified SecureBox SDS2100.
Secure Shell (SSH) is a program that provides strong authentication and secure
communications over unsecured channels. It is used as a replacement for Telnet, rlogin, rsh,
and rcp, to log into another computer over a network, to execute commands in a remote
machine, and to move files from one machine to another. AES is one of the many encryption
algorithms supported by SSH. Once a session key is established SSH uses AES to protect
data in transit.
Both SSH and AES are extremely important to overall network security by maintaining strict
authentication for protection against intruders as well as symmetric encryption to protect
transmission of dangerous packets. AES certification is reliable and can be trusted to handle
the highest network security issues.
WEP
Wired Equivalent Privacy (WEP) is a security protocol for wireless local area networks
(WLANs) which are defined in the 802.11b standard. WEP is designed to provide the same
level of security as that of a wired LAN, however LANs provide more security by their
inherent physical structure that can be protected from unauthorized access. WLANs, which
are over radio waves, do not have the same physical structure and therefore are more
vulnerable to tampering. WEP provides security by encrypting data over radio waves so that
it is protected as it is transmitted from one end point to another. However, it has been found
that WEP is not as secure as once believed. WEP is used at the data link and physical layers
of the OSI model and does not offer end-to-end security.
WPA
Supported by many newer devices, Wi-Fi Protected Access (WPA) is a Wi-Fi standard that
was designed to improve upon the security features of WEP. WPA technology works with
existing Wi-Fi products that have been enabled with WEP, but WPA includes two
improvements over WEP. The first is improved data encryption via the temporal key integrity
protocol (TKIP), which scrambles keys using a hashing algorithm and adds an integrity-
checking feature to ensure that keys havent been tampered with. The second is user
authentication through the extensible authentication protocol (EAP). EAP is built on a secure
public-key encryption system, ensuring that only authorized network users have access. EAP
is generally missing from WEP, which regulates access to a wireless network based on the
computers hardware-specific MAC Address. Since this information can be easily stolen,
there is an inherent security risk in relying on WEP encryption alone.
Applications Abound
While this paper provides a quick snapshot of device servers at work in a variety of
applications, it should be noted that this is only a sampling of the many markets where these
devices could be used. With the ever-increasing requirement to manage, monitor, diagnose
and control many and different forms of equipment and as device server technology
continues to evolve, the applications are literally only limited by the imagination.
Glossary of terms *
Serial server traditionally, a unit used for connecting a modem to the network for
shared access among users.
Terminal server traditionally, a unit that connects asynchronous devices such as
terminals, printers, hosts, and modems to a LAN or WAN.
Device server a specialized network-based hardware device designed to perform a
single or specialized set of functions with client access independent of any operating
system or proprietary protocol.
Print server a host device that connects and manages shared printers over a network.
Console server software that allows the user to connect consoles from various
equipment into the serial ports of a single device and gain access to these consoles
from anywhere on the network.
Console manager a unit or program that allows the user to remotely manage serial
devices, including servers, switches, routers and telecom equipment.
Fast Ethernet Tutorial
A Guide to Using Fast Ethernet and Gigabit Ethernet
Network managers today must contend with the requirements of utilizing faster media,
mounting bandwidth and play traffic cop to an ever-growing network infrastructure. Now,
more than ever, its imperative for them to understand the basics of using various Ethernet
technologies to manage their networks.
This tutorial will explain the basic principles of Fast Ethernet and Gigabit Ethernet
technologies, describing how each improves on basic Ethernet technology. It will offer
guidance on how to implement these technologies as well as some rules of the road for
successful repeater selection and usage.
A good starting point is to explain what Ethernet is. Simply, Ethernet is a very common
method of networking computers in a LAN using copper cabling. Capable of providing fast
and constant connections, Ethernet can handle about 10,000,000 bits per second and can be
used with almost any kind of computer.
While that may sound fast to those less familiar with networking, there is a very strong
demand for even higher transmission speeds, which has been realized by the Fast Ethernet
and Gigabit Ethernet specifications (IEEE 802.3u and IEEE 802.3z respectively). These LAN
(local area network) standards have raised the Ethernet speed limit from 10 megabits per
second (Mbps) to 100Mbps for Fast Ethernet and 1000Mbps for Gigabit Ethernet with only
minimal changes made to the existing cable structure.
The building blocks of todays networks call out for a mixture of legacy 10BASE-T Ethernet
networks and the new protocols. Typically, 10Mbps networks utilize Ethernet switches to
improve the overall efficiency of the Ethernet network. Between Ethernet switches, Fast
Ethernet repeaters are used to connect a group of switches together at the higher 100 Mbps
rate.
However, with an increasing number of users running 100Mbps at the desktop, servers and
aggregation points such as switch stacks may require even greater bandwidth. In this case, a
Fast Ethernet backbone switch can be upgraded to a Gigabit Ethernet switch which supports
multiple 100/1000 Mbps switches. High performance servers can be connected directly to the
backbone once it has been upgraded.
When integrating 100BASE-T into a 10BASE-T network, the only change required from a
wiring standpoint is that the corporate premise distributed wiring system must now include
Category 5 (CAT5) rated twisted pair cable in the areas running 100BASE-T. Once rewiring
is completed, gigabit speeds can also be deployed even more widely throughout the network
using standard CAT5 cabling.
The Fast Ethernet specification calls for two types of transmission schemes over various wire
media. The first is 100BASE-TX, which, from a cabling perspective, is very similar to
10BASE-T. It uses CAT5-rated twisted pair copper cable to connect various hubs, switches
and end-nodes. It also uses an RJ45 jack just like 10BASE-T and the wiring at the connector
is identical. These similarities make 100BASE-TX easier to install and therefore the most
popular form of the Fast Ethernet specification.
The second variation is 100Base-FX which is used primarily to connect hubs and switches
together either between wiring closets or between buildings. 100BASE-FX uses multimode
fiber-optic cable to transport Fast Ethernet traffic.
Gigabit Ethernet specification calls for three types of transmission schemes over various wire
media. Gigabit Ethernet was originally designed as a switched technology and used fiber for
uplinks and connections between buildings. Because of this, in June 1998 the IEEE approved
the Gigabit Ethernet standard over fiber: 1000BASE-LX and 1000BASE-SX.
The next Gigabit Ethernet standardization to come was 1000BASE-T, which is Gigabit
Ethernet over copper. This standard allows one gigabit per second (Gbps) speeds to be
transmitted over CAT5 cable and has made Gigabit Ethernet migration easier and more cost-
effective than ever before.
Class I Repeater The Class 1 repeater operates by translating line signals on the incoming
port to a digital signal. This allows the translation between different types of Fast Ethernet
such as 100BASE-TX and 100BASE-FX. A Class I repeater introduces delays when
performing this conversion such that only one repeater can be put in a single Fast Ethernet
LAN segment.
Class II Repeater The Class II repeater immediately repeats the signal on an incoming
port to all the ports on the repeater. Very little delay is introduced by this quick movement of
data across the repeater; thus two Class II repeaters are allowed per Fast Ethernet segment.
Network managers understand the 100 meter distance limitation of 10BASE-T and
100BASE-T Ethernet and make allowances for working within these limitations. At the
higher operating speeds, Fast Ethernet and 1000BASE-T are limited to 100 meters over
CAT5-rated cable. The EIA/TIA cabling standard recommends using no more than 90 meters
between the equipment in the wiring closet and the wall connector. This allows another 10
meters for patch cables between the wall and the desktop computer.
In contrast, a Fast Ethernet network using the 100BASE-FX standard is designed to allow
LAN segments up to 412 meters in length. Even though fiber-optic cable can actually
transmit data greater distances (i.e. 2 Kilometers in FDDI), the 412 meter limit for Fast
Ethernet was created to allow for the round trip times of packet transmission. Typical
100BASE-FX cable specifications call for multimode fiber-optic cable with a 62.5 micron
fiber-optic core and a 125 micron cladding around the outside. This is the most popular fiber
optic cable type used by many of the LAN standards today. Connectors for 100BASE-FX
Fast Ethernet are typically ST connectors (which look like Ethernet BNC connectors).
Many Fast Ethernet vendors are migrating to the newer SC connectors used for ATM over
fiber. A rough implementation guideline to use when determining the maximum distances in
a Fast Ethernet network is the equation: 400 (r x 95) where r is the number of repeaters.
Network managers need to take into account the distance between the repeaters and the
distance between each node from the repeater. For example, in Figure 1 two repeaters are
connected to two Fast Ethernet switches and a few servers.
There is yet another variation of Ethernet called full-duplex Ethernet. Full-duplex Ethernet
enables the connection speed to be doubled by simply adding another pair of wires and
removing collision detection; the Fast Ethernet standard allowed full-duplex Ethernet. Until
then all Ethernet worked in half-duplex mode which meant if there were only two stations on
a segment, both could not transmit simultaneously. With full-duplex operation, this was now
possible. In the terms of Fast Ethernet, essentially 200Mbps of throughput is the theoretical
maximum per full-duplex Fast Ethernet connection. This type of connection is limited to a
node-to-node connection and is typically used to link two Ethernet switches together.
A Gigabit Ethernet network using the 1000BASE-LX long wavelength option supports
duplex links of up to 550 meters of 62.5 millimeters or 50 millimeters multimode fiber.
1000BASE-LX can also support up to 5 Kilometers of 10 millimeter single-mode fiber. Its
wavelengths range from 1270 millimeters to 1355 millimeters. The 1000BASE-SX is a short
wavelength option that supports duplex links of up to 275 meters using 62.5 millimeters at
multimode or up to 550 meters using 55 millimeters of multimode fiber. Typical wavelengths
for this option are in the range of 770 to 860 nanometers.
The basic strategy of cabling Fast Ethernet systems is to minimize the re-transmission of
packets caused by high bit-error rates. This ratio is calculated using NEXT, ambient noise
and attenuation of the cable.
Switches can be a valuable asset to networking. Overall, they can increase the capacity and
speed of your network. However, switching should not be seen as a cure-all for network
issues. Before incorporating network switching, you must first ask yourself two important
questions: First, how can you tell if your network will benefit from switching? Second, how
do you add switches to your network design to provide the most benefit?
This tutorial is written to answer these questions. Along the way, well describe how switches
work, and how they can both harm and benefit your networking strategy. Well also discuss
different network types, so you can profile your network and gauge the potential benefit of
network switching for your environment.
What is a Switch?
Switches occupy the same place in the network as hubs. Unlike hubs, switches examine each
packet and process it accordingly rather than simply repeating the signal to all ports. Switches
map the Ethernet addresses of the nodes residing on each network segment and then allow
only the necessary traffic to pass through the switch. When a packet is received by the switch,
the switch examines the destination and source hardware addresses and compares them to a
table of network segments and addresses. If the segments are the same, the packet is dropped
or filtered; if the segments are different, then the packet is forwarded to the proper
segment. Additionally, switches prevent bad or misaligned packets from spreading by not
forwarding them.
Filtering packets and regenerating forwarded packets enables switching technology to split a
network into separate collision domains. The regeneration of packets allows for greater
distances and more nodes to be used in the total network design, and dramatically lowers the
overall collision rates. In switched networks, each segment is an independent collision
domain. This also allows for parallelism, meaning up to one-half of the computers connected
to a switch can send data at the same time. In shared networks all nodes reside in a single
shared collision domain.
Easy to install, most switches are self learning. They determine the Ethernet addresses in use
on each segment, building a table as packets are passed through the switch. This plug and
play element makes switches an attractive alternative to hubs.
Switches can connect different network types (such as Ethernet and Fast Ethernet) or
networks of the same type. Many switches today offer high-speed links, like Fast Ethernet,
which can be used to link the switches together or to give added bandwidth to important
servers that get a lot of traffic. A network composed of a number of switches linked together
via these fast uplinks is called a collapsed backbone network.
Dedicating ports on switches to individual nodes is another way to speed access for critical
computers. Servers and power users can take advantage of a full segment for one node, so
some networks connect high traffic nodes to a dedicated switch port.
Network Congestion
As more users are added to a shared network or as applications requiring more data are
added, performance deteriorates. This is because all users on a shared network are
competitors for the Ethernet bus. A moderately loaded 10 Mbps Ethernet network is able to
sustain utilization of 35 percent and throughput in the neighborhood of 2.5 Mbps after
accounting for packet overhead, inter-packet gaps and collisions. A moderately loaded Fast
Ethernet or Gigabit Ethernet shares 25 Mbps or 250 Mbps of real data in the same
circumstances. With shared Ethernet and Fast Ethernet, the likelihood of collisions increases
as more nodes and/or more traffic is added to the shared collision domain.
Ethernet itself is a shared media, so there are rules for sending packets to avoid conflicts and
protect data integrity. Nodes on an Ethernet network send packets when they determine the
network is not in use. It is possible that two nodes at different locations could try to send data
at the same time. When both PCs are transferring a packet to the network at the same time, a
collision will result. Both packets are retransmitted, adding to the traffic problem.
Minimizing collisions is a crucial element in the design and operation of networks. Increased
collisions are often the result of too many users or too much traffic on the network, which
results in a great deal of contention for network bandwidth. This can slow the performance of
the network from the users point of view. Segmenting, where a network is divided into
different pieces joined together logically with switches or routers, reduces congestion in an
overcrowded network by eliminating the shared collision domain.
Collision rates measure the percentage of packets that are collisions. Some collisions are
inevitable, with less than 10 percent common in well-running networks.
Utilization rate is another widely accessible statistic about the health of a network. This
statistic is available in Novells console monitor and WindowsNT performance monitor as
well as any optional LAN analysis software. Utilization in an average network above 35
percent indicates potential problems. This 35 percent utilization is near optimum, but some
networks experience higher or lower utilization optimums due to factors such as packet size
and peak load deviation.
A switch is said to work at wire speed if it has enough processing power to handle full
Ethernet speed at minimum packet sizes. Most switches on the market are well ahead of
network traffic capabilities supporting the full wire speed of Ethernet, 14,480 pps (packets
per second), and Fast Ethernet, 148,800 pps.
Routers
Routers work in a manner similar to switches and bridges in that they filter out network
traffic. Rather than doing so by packet addresses, they filter by specific protocol. Routers
were born out of the necessity for dividing networks logically instead of physically. An IP
router can divide a network into various subnets so that only traffic destined for particular IP
addresses can pass between segments. Routers recalculate the checksum, and rewrite the
MAC header of every packet. The price paid for this type of intelligent forwarding and
filtering is usually calculated in terms of latency, or the delay that a packet experiences inside
the router. Such filtering takes more time than that exercised in a switch or bridge which only
looks at the Ethernet address. In more complex networks network efficiency can be
improved. An additional benefit of routers is their automatic filtering of broadcasts, but
overall they are complicated to setup.
Switch Benefits
Isolates traffic, relieving congestion
Separates collision domains, reducing collisions
Segments, restarting distance and repeater rules
Switch Costs
Price: currently 3 to 5 times the price of a hub
Packet processing time is longer than in a hub
Monitoring the network is more complicated
Since switches are self learning, they are as easy to install as a hub. Just plug them in and go.
And they operate on the same hardware layer as a hub, so there are no protocol issues.
There are two reasons for switches being included in network designs. First, a switch breaks
one network into many small networks so the distance and repeater limitations are restarted.
Second, this same segmentation isolates traffic and reduces collisions relieving network
congestion. It is very easy to identify the need for distance and repeater extension, and to
understand this benefit of network switching. But the second benefit, relieving network
congestion, is hard to identify and harder to understand the degree by which switches will
help performance. Since all switches add small latency delays to packet processing,
deploying switches unnecessarily can actually slow down network performance. So the next
section pertains to the factors affecting the impact of switching to congested networks.
Network Switching
The benefits of switching vary from network to network. Adding a switch for the first time
has different implications than increasing the number of switched ports already installed.
Understanding traffic patterns is very important to network switching the goal being to
eliminate (or filter) as much traffic as possible. A switch installed in a location where it
forwards almost all the traffic it receives will help much less than one that filters most of the
traffic.
Networks that are not congested can actually be negatively impacted by adding switches.
Packet processing delays, switch buffer limitations, and the retransmissions that can result
sometimes slows performance compared with the hub based alternative. If your network is
not congested, dont replace hubs with switches. How can you tell if performance problems
are the result of network congestion? Measure utilization factors and collision rates.
Network response times (the user-visible part of network performance) suffers as the load on
the network increases, and under heavy loads small increases in user traffic often results in
significant decreases in performance. This is similar to automobile freeway dynamics, in that
increasing loads results in increasing throughput up to a point, then further increases in
demand results in rapid deterioration of true throughput. In Ethernet, collisions increase as
the network is loaded, and this causes retransmissions and increases in load which cause even
more collisions. The resulting network overload slows traffic considerably.
Using network utilities found on most server operating systems network managers can
determine utilization and collision rates. Both peak and average statistics should be
considered.
Replacing a Central Hub with a Switch
This switching opportunity is typified by a fully shared network, where many users are
connected in a cascading hub architecture. The two main impacts of switching will be faster
network connection to the server(s) and the isolation of non-relevant traffic from each
segment. As the network bottleneck is eliminated performance grows until a new system
bottleneck is encountered such as maximum server performance.
Fast Ethernet is very easy to add to most networks. A switch or bridge allows Fast Ethernet to
connect to existing Ethernet infrastructures to bring speed to critical links. The faster
technology is used to connect switches to each other, and to switched or shared servers to
ensure the avoidance of bottlenecks.
Many client/server networks suffer from too many clients trying to access the same server
which creates a bottleneck where the server attaches to the LAN. Fast Ethernet, in
combination with switched Ethernet, creates the perfect cost-effective solution for avoiding
slow client server networks by allowing the server to be placed on a fast port.
Distributed processing also benefits from Fast Ethernet and switching. Segmentation of the
network via switches brings big performance boosts to distributed traffic networks, and the
switches are commonly connected via a Fast Ethernet backbone.
Managed or Unmanaged
Management provides benefits in many networks. Large networks with mission critical
applications are managed with many sophisticated tools, using SNMP to monitor the health
of devices on the network. Networks using SNMP or RMON (an extension to SNMP that
provides much more data while using less network bandwidth to do so) will either manage
every device, or just the more critical areas. VLANs are another benefit to management in a
switch. A VLAN allows the network to group nodes into logical LANs that behave as one
network, regardless of physical connections. The main benefit is managing broadcast and
multicast traffic. An unmanaged switch will pass broadcast and multicast packets through to
all ports. If the network has logical grouping that are different from physical groupings then a
VLAN-based switch may be the best bet for traffic optimization.
Another benefit to management in the switches is Spanning Tree Algorithm. Spanning Tree
allows the network manager to design in redundant links, with switches attached in loops.
This would defeat the self learning aspect of switches, since traffic from one node would
appear to originate on different ports. Spanning Tree is a protocol that allows the switches to
coordinate with each other so that traffic is only carried on one of the redundant links (unless
there is a failure, then the backup link is automatically activated). Network managers with
switches deployed in critical applications may want to have redundant links. In this case
management is necessary. But for the rest of the networks an unmanaged switch would do
quite well, and is much less expensive.
Today, the speed of store-and-forward switches has caught up with cut-through switches to
the point where the difference between the two is minimal. Also, there are a large number of
hybrid switches available that mix both cut-through and store-and-forward architectures.
Take a switchs specifications and add up all the ports at theoretical maximum speed, then
you have the theoretical sum total of a switchs throughput. If the switching bus, or switching
components cannot handle the theoretical total of all ports the switch is considered a
blocking switch. There is debate whether all switches should be designed non-blocking, but
the added costs of doing so are only reasonable on switches designed to work in the largest
network backbones. For almost all applications, a blocking switch that has an acceptable and
reasonable throughput level will work just fine.
Consider an eight port 10/100 switch. Since each port can theoretically handle 200 Mbps (full
duplex) there is a theoretical need for 1600 Mbps, or 1.6 Gbps. But in the real world each
port will not exceed 50% utilization, so a 800 Mbps switching bus is adequate. Consideration
of total throughput versus total ports demand in the real world loads provides validation that
the switch can handle the loads of your network.
As packets are processed in the switch, they are held in buffers. If the destination segment is
congested, the switch holds on to the packet as it waits for bandwidth to become available on
the crowded segment. Buffers that are full present a problem. So some analysis of the buffer
sizes and strategies for handling overflows is of interest for the technically inclined network
designer.
In real world networks, crowded segments cause many problems, so their impact on switch
consideration is not important for most users, since networks should be designed to eliminate
crowded, congested segments. There are two strategies for handling full buffers. One is
backpressure flow control which sends packets back upstream to the source nodes of
packets that find a full buffer. This compares to the strategy of simply dropping the packet,
and relying on the integrity features in networks to retransmit automatically. One solution
spreads the problem in one segment to other segments, propagating the problem. The other
solution causes retransmissions, and that resulting increase in load is not optimal. Neither
strategy solves the problem, so switch vendors use large buffers and advise network
managers to design switched network topologies to eliminate the source of the problem
congested segments.
Layer 3 Switching
A hybrid device is the latest improvement in internetworking technology. Combining the
packet handling of routers and the speed of switching, these multilayer switches operate on
both layer 2 and layer 3 of the OSI network model. The performance of this class of switch is
aimed at the core of large enterprise networks. Sometimes called routing switches or IP
switches, multilayer switches look for common traffic flows, and switch these flows on the
hardware layer for speed. For traffic outside the normal flows, the multilayer switch uses
routing functions. This keeps the higher overhead routing functions only where it is needed,
and strives for the best handling strategy for each network packet.
Many vendors are working on high end multilayer switches, and the technology is definitely
a work in process. As networking technology evolves, multilayer switches are likely to
replace routers in most large networks.