SlideShare a Scribd company logo
Introduction to
                         Fibre Channel over Ethernet (FCoE)
                                                                             A Detailed Review




Abstract
Fibre Channel over Ethernet (FCoE) is a new storage networking option, based on industry standards,
which is now being deployed in real world environments. This white paper provides an overview of FCoE,
describes the hardware and software components that make up the ecosystem, and explains how the
technology is expected to continue to mature over the next few years.


                                                                                            April 2011
Copyright © 2008, 2009, 2010, 2011 EMC Corporation. All rights reserved.
EMC believes the information in this publication is accurate as of its publication date. The information is
subject to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED ―AS IS.‖ EMC CORPORATION
MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE
INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED
WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicable
software license.
For the most up-to-date listing of EMC product names, see EMC Corporation Trademarks on EMC.com
VMware and ESX are registered trademarks or trademarks of VMware, Inc. in the United States and/or
other jurisdictions. All other trademarks used herein are the property of their respective owners.
Part Number h5916.3

Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                             2
Table of Contents

Executive summary ............................................................................................ 4
Introduction ......................................................................................................... 5
  Audience ...................................................................................................................................... 6
Creation of an FCoE infrastructure ................................................................... 6
FCoE has FC inside ............................................................................................ 6
FCoE is FC over ‘Lossless Ethernet’ ................................................................ 7
Road to a converged fabric ................................................................................ 8
Conclusion ........................................................................................................ 10




Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                                                                3
Executive summary
The question for most data center managers is not should they network their storage, but which of the many
options available should they use and for which applications can they justify the expense. Using networked
storage reduces not only capital expenses through higher utilization of resources and the ability to create
highly available configurations, but also operational expenses by allowing centralized management and
simplified backup and replication. While most companies are using networked storage, overall less than 20
percent1 of all data center servers are attached to networked storage. Virtualized servers have over an 80
percent2 network attach rate since many advanced features such as mobility and high availability cannot be
utilized if these servers are not using networked storage. It is expected that as the overall trend of server
virtualization increases that the percentage of servers attached to networked storage in the data center will
increase.
One of the reasons why more servers do not utilize networked storage is that the storage network (Fibre
Channel) is separate from the Ethernet (TCP/IP) network (see Figure 1). Fibre Channel is the predominate
choice for networked storage in the data center because it allows for the creation of a highly scalable –
hundreds or thousands of nodes – and reliable solution that can be centrally managed. For networking,
today’s servers typically have multiple 1 Gb Ethernet network interface cards (NICs). The number of NICs
in a system will vary depending on the applications on the machine, and management and security
requirements, but the number can be four, six, eight, or more NICs in a single machine. This high number
of server adapters, cables, and separate networking and storage fabrics adds to complexity and cost. As a
result, many customers would like to be able to have all applications run over a single converged network.
This goal of I/O consolidation on a unified network for all traffic leads to a savings in infrastructure
(cabling, reduced sparing, and so on) and a simplified management environment. Overall data center
trends require solutions that take up less space, draw less power, and require less cooling; suppliers are
reaching a breaking point where supporting multiple separate networks will not allow them to meet these
requirements.




Figure 1. Existing data center with separate Ethernet and FC networks

1
  IDC, Worldwide Fibre Channel Host Bus Adapter 2007–2011 Forecast and 2006 Vendor Shares and
Worldwide and Regional Server 2007-2011 Forecast
2
  Enterprise Storage Group, Server Virtualization: The Impact on Storage, November 7, 2007

Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                           4
Ethernet has been given top consideration as a potential solution for I/O consolidation due to the large
install base and broad general understanding of the technology by all IT departments. While some storage
applications can run comfortably with 1 Gb bandwidth widely available with Ethernet, many data center
solutions require the higher bandwidth solutions that Fibre Channel can supply today with 2 Gb, 4 Gb, 8
Gb, and 16 Gb options. With the maturation of 10 Gb Ethernet, however, there is an inflection point where
we have the opportunity to consolidate the many existing connections in a server to a pair of 10 Gb
Ethernet links. Blade servers and server virtualization solutions, as well as newer server bus architectures
like PCI Express 2.0, will allow infrastructure to utilize the bandwidth of 10 Gb adapters. The challenge is
in unifying storage traffic and network traffic onto the same link. Fibre Channel over Ethernet (FCoE) will
allow an Ethernet-based SAN to be introduced into the FC-based data center without breaking existing
administration tools or workflows. This is done by allowing Fibre Channel traffic to run over an Ethernet
infrastructure. Servers and storage see FCoE as FC (Figure 2) since FCoE is simply encapsulating FC over
the Ethernet portions of the connectivity, while to the server applications as well as FC SAN and FC
storage, the traffic is still FC. FCoE can be easily added into existing FC environments in contrast to
previous attempts at consolidation with iSCSI and InfiniBand, which required new drivers and a rip-and-
replace for customers that are using Fibre Channel. The security and management best practices of having
a single application per NIC will need to be revisited as part of 10 Gb Ethernet adoption.



                                                                         Ethernet
     Network                   FC                                        Network
      Driver                  Driver



      Converged Network Adapter
                                Encapsulate
                                                                                             FC storage

            Lossless Ethernet Links
                                                                    FC network

                            FCoE
                            Switch

      Ethernet          De-Encapsulate
      FC


Figure 2. FCoE environment with converged adapters at the server attached to a
converged switch


Introduction
Today’s networks use different protocols to send information between devices. Traditional Ethernet is a
family of frame-based computer networking technologies for local area networks (LANs), whereas Fibre
Channel is used for storage area networking (SANs). Fibre Channel over Ethernet, or FCoE, is a new
storage networking protocol that supports Fibre Channel natively over Ethernet. FCoE encapsulates Fibre
Channel frames into Ethernet frames, allowing them to run alongside traditional Internet Protocol (IP)
traffic. This white paper provides an overview of FCoE, describes the hardware and software components
that make up the new ecosystem, and explains how the technology is expected to mature over the next few
years.



Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                          5
Audience
This white paper is intended for network and storage administrators who want to learn more about FCoE
and its benefits.


Creation of an FCoE infrastructure
FCoE requires the deployment of three new components: a Converged Network Adapter (CNA), Lossless
Ethernet Links, and an FCoE switch. The CNA provides the functions of both a standard NIC and a FC
HBA in a single adapter in the server. There are two types of CNAs available: a ―hardware‖-based solution
— where the lower-level FC (exchange and sequence management) and FCoE functions are done in
hardware — or a ―software‖ solution — where some of those functions are performed in software. When
obtaining a hardware solution from a vendor that manufacturers FC HBAs, the drivers will be the same for
both FC and FCoE adapters; this provides the least disruption and easiest integration into existing
environments. The software solutions require the development of new FC drivers that will be run in
software; one of the efforts can be found at Open-FCoE.org. The FCoE switch is the network device that
will connect to existing LAN and SAN environments. The T11 standards FC-BB-5 group
(http://www.t11.org/fcoe) created the FCoE protocol, which enables the creation of CNAs and FCoE
switches. The link that FCoE traffic uses must provide the same level of lossless behavior that can be
found with Fibre Channel environments today. The Data Center Bridging group (part of the IEEE 802.1
standards, see http://www.ieee802.org/1/pages/dcbridges.html) is investigating requirements to create an
Ethernet environment that can support storage and IPC traffic. The enhancements to Ethernet will be
discussed later, but new functionality is required in the devices and 10 Gb Ethernet is going to be the entry-
level solution. Customers have started to introduce 10 Gb Ethernet in their data centers, so the FCoE effort
is poised to help mold the technology and catch the incoming wave of adoption.



FCoE has FC inside
In creating the FCoE protocol, the goal is summed up in the name — take FC and put it over an Ethernet
infrastructure. To do this, the Fibre Channel frames are encapsulated — not translated or bridged — in an
Ethernet frame (Figure 3). The mapping is 1:1, which means that there is no segmenting of FC frames nor
are multiple FC frames put in a single Ethernet frame. It is a requirement that all devices for FCoE
(adapters and switches along the path) support mini-jumbo frames to allow the largest FC frame to be
supported without segmentation. Flow control of FCoE is controlled by Priority Flow Control mechanisms
rather than by the buffer credits of FC. The naming conventions for FCoE port types are very similar to
FC:
     N_Ports (HBAs and storage) are VN_Ports (CNAs or FCoE storage).
    F_Ports (fabric ports) are VF_Ports (FCoE switch ports that attach to VN_Ports).
    E_Ports (switch to switch) are VE_Ports (ports between two FCoE Ethernet switches).




Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                           6
Figure 3. FCoE frame format

In Fibre Channel, security is typically not high on the list of discussion points. In the early days of FC, the
small-scale physical security and optical cabling were usually sufficient for customers. The Fibre Channel
Security Protocols standard (FC-SP) was approved in 2004 and addresses how to protect against security
breaches. Ethernet has its own well-established security practices, so the standards were only concerned
with specific threats that are not covered in a typical Ethernet environment. An issue of concern was that
while FC links are always private (two ends), FCoE links can be shared (three-plus ends) due to the nature
of Layer 2 Ethernet. Access Control Lists (ACLs) will be used to specify on a port basis what operations
are allowed to be performed, similar to the protection provided by a firewall. A feature called ―Dynamic
ACLs‖ is supported by the FCoE switches to reduce the administrative burden of manually configuring
ACLs.
FC-BB-5, the T11 standard for FCoE was ratified in June 2009.


FCoE is FC over ‘Lossless Ethernet’
The Ethernet infrastructure, over which FCoE will travel, must be of a lossless nature. Since the FCoE
protocol does not contain TCP, any lost packets would require recovery at the SCSI layer. In a typical
environment, based on existing Ethernet, this would happen much too often to be usable for storage
environments. Fortunately, the Ethernet community has been looking at the issue of creating a lossless
environment for a number of years. The set of features required to support lossless behavior is referred to
as Data Center Bridging (DCB) and is being worked on by the IEEE DCB working group. The first set of
features is link level protocols and by definition only covers the link between two devices (that is, either an
adapter and switch port or two switch ports); they are Priority Flow Control (PFC) and Enhanced
Transmission Selection (ETS).
Even before FCoE was conceived, there was a PAUSE function (Link Pause, IEEE 802.3, Annex 31B) that
would allow for the creation of a lossless environment by stopping all traffic on a port when a full queue
condition was achieved. The problem with classical PAUSE is that you cannot specify which traffic should
be lossless, and as a result there is no mechanism to deal with the congestion created on all classes of traffic
when a PAUSE is issued. Priority Flow Control (IEEE 802.1Qbb) creates a new PAUSE function that can
halt traffic according to the priority tag while allowing traffic at other priority levels to continue.



Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                                 7
Administrators use the eight lanes defined in IEEE 802.1p to create virtual lossless lanes for traffic classes
like storage (that require lossless behavior) and lossy lanes for other classes.
In order to make sure that one traffic stream does not take too much of the overall bandwidth, the addition
of Enhanced Transmission Selection (IEEE 802.1Qaz) is important for creating an environment where
storage and other traffic can share the same link. A common management framework can be created for
bandwidth management. High-priority traffic like storage can be prioritized and allocated bandwidth while
still allowing other traffic classes to utilize the bandwidth when it is not being fully utilized.
The Data Center Bridging Exchange Protocol (DCBX) is responsible for the configuration of link
parameters for Data Center Bridging functions. It determines which devices support the enhanced
functionalities that create the ―DCB‖ cloud where FCoE traffic can safely travel. It also allows for
transparent pass-through for non-DCB traffic classes.
These link level enhancements start to create a safe
environment for FCoE traffic. PFC, ETS, and DCBX have                  Virtualization and FCoE
all been defined and therefore solutions that have these          Currently, server virtualization
functions shipping today should be able to support FCoE.          environments (including VMware and
                                                                  Microsoft Hyper-V) can choose among
As the configurations expand, there are additional Ethernet       FC, iSCSI, and NAS for networked
enhancements beyond link level that will increase                 storage. The hypervisor has a storage
performance. Congestion Notification (IEEE 802.1Qau) is           driver stack that presents FC (from an
a development to manage congestion beyond the link level.         HBA) and iSCSI (from a NIC) traffic to
This is desirable when a link is reaching a PAUSE                 individual virtual machines (VMs) or
condition and the source of the traffic can be alerted and        guests as storage. There is also a virtual
throttled appropriately, reducing congestion out of the           switch in the hypervisor that sends
network core. While not necessary to support FCoE, an             traditional LAN traffic to the VM/guest as
additional enhancement to optimize Ethernet for storage           network traffic. FCoE solutions using a
environments is Layer 2 multipathing. Ethernet solutions          CNA will function equivalently as
today use Spanning Tree Protocol (STP), which only                existing solutions – the CNA will convert
supports a single active path between any two network             FCoE traffic to FC packets in the
nodes. The IETF standards group has created TRILL                 hardware, so the hypervisor and VM/guest
(TRansparent Interconnection of Lots of Links), which             will still work as if physical FC HBAs and
allows for increased bandwidth by allowing and                    NICs were installed. Software FCoE
aggregating multiple network paths.                               solutions with a standard NIC will require
                                                                  additional developments before they can
                                                                  be used. Currently hypervisors will not
Road to a converged fabric                                        identify FCoE traffic that is still
Many customers are already running their storage traffic          encapsulated and virtual switches
over Ethernet by utilizing iSCSI or NAS; FCoE is not              (including the Cisco Nexus 1000V in
designed to replace these solutions. For customers that           VMware® ESX®) do not have Lossless
have an existing infrastructure and knowledge base of FC,         Ethernet functionality, so sending traffic
FCoE provides a path toward reaching a converged fabric.          directly to the VM/guest would not be
FCoE, iSCSI, and NAS will all leverage 10 Gigabit                 reliable. FCoE will expand the available
Ethernet. In general, iSCSI environments tend to be small         storage networking solutions options for
configurations with under 100 servers, while most FC              the high-growth server virtualization
customers are scaling into hundreds or thousands of nodes;        market. Moving to a converged fabric
FCoE can be plugged in to these existing environments.            will simplify mobility by reducing the
10 Gigabit Ethernet will provide the common                       number of networks that need to be
infrastructure that can support FCoE, iSCSI, and NAS              moved. It also allows a standard
(Figure 4), all of which can be used in virtualization            configuration to be built, allowing for
environments, and customers will choose based upon their          flexible storage deployment.
requirements and skill set.




Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                                8
10 Gigabit
                Ethernet                                  Fibre
                                                         Channel


                                             common management
            common infrastructure




                                virtualization

Figure 4. 10 Gigabit Ethernet common infrastructure for storage
All new technologies take time to develop a robust ecosystem, and FCoE is no exception. Since FCoE
builds on the existing infrastructure, it can be rolled out in a phased manner. Today FCoE can support
convergence at the server, in which the separate network (NIC) and storage (HBA) interfaces on the server
are replaced by the CNA that is directly attached to the FCoE switch. A new rack of servers can deploy
these technologies, while the existing storage and network environments remain unchanged. FCoE can
also support convergence at the storage port and on switch-to-switch links, allowing for an end-to-end
(FCoE multi-hop) configuration to be created.
Recently, FCoE solutions started to expand beyond a single switch at the access layer into multi-hop
configurations. While the industry continues to work on the technical issues of expanding FCoE solutions,
customers should review the structure of their storage and networking groups to make sure that there are
not any ―political‖ issues over who owns the purchase, installation, and maintenance of the CNA and FCoE
switch components in a converged network environment. Some of the enhancements to Ethernet will be
new to network administrators who will need to learn these functions and will require coordination with
storage administrators. The overlapping domains may compel cultural adjustments, as storage networks
will no longer be dedicated and network configurations can no longer be reconfigured at-will. Network and
storage administrators will each have their own management interfaces to the environments, keeping tasks
separate rather than converged.
As director-class products that support FCoE are released into the market, building FCoE multi-hop
configurations that include native FCoE storage begins to make sense.
No discussion of FCoE is complete without addressing the cabling infrastructure supporting the solution.
When customers deploy a physical cabling infrastructure, they are looking for solutions that can be used for
five to ten years. For 1 Gigabit Ethernet, the primary options for cabling are copper (1000Base-T with RJ-
45 connectors) and optical (same physical cabling as FC); copper dominates this market with billions of
ports installed while optical has historically been 1 percent of Ethernet ports3. A standard (10GBase-T) for
using existing copper cabling (either CAT 6 or CAT 6a with RJ-45 connectors) is available but not
expected to start shipping for FCoE solutions until 2012. A new copper option known as Twinax has
become available for FCoE solutions of 10 Gigabit Ethernet. The option is based on the SFF-8431
standard and uses the SFP+ interface for a copper connection that is low cost and low power. The Twinax
solution is limited to short distances, supporting between 1-10 meters, which is sufficient for server to top-


3
    IEEE Installed Cabling market data (http://ieee802.org/3/10GBT/public/jan03/flatman_1_0103.pdf)

Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                            9
of-rack or end-of-row switch environments. Standard multimode optical cabling will be used for
environments that require longer distances such as from the rack to the core.


Conclusion
FCoE is a natural extension of FC networks that will allow data centers to leverage converged I/O and
extend the operational efficiencies of networked storage to a new wave of servers. Server, storage, and
networking vendors have all invested heavily to create converged fabric technologies. There is a robust
FCoE ecosystem that will continue to expand and mature over time.




Introduction to Fibre Channel over Ethernet (FCoE)
A Detailed Review                                                                                         10

More Related Content

PDF
Converged Data Center: FCoE, iSCSI, & the Future of Storage Networking ( EMC ...
 
PPTX
VMware EMC Service Talk
PDF
FC/FCoE - Topologies, Protocols, and Limitations ( EMC World 2012 )
 
PPT
Converged Networks: FCoE, iSCSI and the Future of Storage Networking
PDF
Converged Data Center: FCoE, iSCSI and the Future of Storage Networking
 
PDF
Fiber Channel over Ethernet (FCoE) – Design, operations and management best p...
PDF
Converged Data Center: FCoE, iSCSI, and the Future of Storage Networking ( EM...
 
PDF
Converged data center_f_co_e_iscsi_future_storage_networking
 
Converged Data Center: FCoE, iSCSI, & the Future of Storage Networking ( EMC ...
 
VMware EMC Service Talk
FC/FCoE - Topologies, Protocols, and Limitations ( EMC World 2012 )
 
Converged Networks: FCoE, iSCSI and the Future of Storage Networking
Converged Data Center: FCoE, iSCSI and the Future of Storage Networking
 
Fiber Channel over Ethernet (FCoE) – Design, operations and management best p...
Converged Data Center: FCoE, iSCSI, and the Future of Storage Networking ( EM...
 
Converged data center_f_co_e_iscsi_future_storage_networking
 

What's hot (18)

PPT
Fibre Channel over Ethernet (FCoE), iSCSI and the Converged Data Center
PDF
Converged Data Center: FCoE, iSCSI and the Future of Storage Networking
 
PPT
Mpls vpn toi
PPTX
Answers to Your IT Nightmares - SAS, iSCSI, or Fibre Channel?
PDF
Mpls concepts. Time to Certify
PDF
MPLS-based Metro Ethernet Networks Tutorial by Khatri
PDF
NetApp Multi-Protocol Storage Evaluation
PDF
Sakar jain
PDF
PLNOG 8: Piotr Głaska - Data Center Networking - Trends and suggestions
PPTX
Maximize Server Refresh Revenue with Emulex 8GB Fibre Channel and the Emulex ...
PDF
Network Virtualization using Shortest Path Bridging
PDF
F co e_netapp_v12
PDF
Mellanox hpc day 2011 kiev
PDF
Webcast: Reduce Costs, Improve Agility with Convergenomics
PDF
MPLS-based Metro Ethernet Networks
PDF
Linkage aggregation
PDF
Shortest path bridging 802.1 aq
PDF
White Paper: IP VPN and Ethernet WAN Services
Fibre Channel over Ethernet (FCoE), iSCSI and the Converged Data Center
Converged Data Center: FCoE, iSCSI and the Future of Storage Networking
 
Mpls vpn toi
Answers to Your IT Nightmares - SAS, iSCSI, or Fibre Channel?
Mpls concepts. Time to Certify
MPLS-based Metro Ethernet Networks Tutorial by Khatri
NetApp Multi-Protocol Storage Evaluation
Sakar jain
PLNOG 8: Piotr Głaska - Data Center Networking - Trends and suggestions
Maximize Server Refresh Revenue with Emulex 8GB Fibre Channel and the Emulex ...
Network Virtualization using Shortest Path Bridging
F co e_netapp_v12
Mellanox hpc day 2011 kiev
Webcast: Reduce Costs, Improve Agility with Convergenomics
MPLS-based Metro Ethernet Networks
Linkage aggregation
Shortest path bridging 802.1 aq
White Paper: IP VPN and Ethernet WAN Services
Ad

Viewers also liked (20)

PDF
Mit2 092 f09_lec23
PPTX
Discover Great Reasons to move to ConfigMgr 2012 SP1
PDF
Lessico civico. Teorie e pratiche della cittadinanza
PPT
Japan russio japanese war
PPTX
Becerrajavier a1
PPTX
elasticity 2014
PPTX
How Does Long-term Care Insurance Work?
PPTX
Day 3 mon world
PPTX
What's new for Lync 2013 Clients & Devices
PPTX
Wed thurs reform
PDF
TechBook: IMS on z/OS Using EMC Symmetrix Storage Systems
 
PPTX
Mongol oron neelttei hicheel
PDF
Protectora d'animals_Xènia, Malina i Gemma
PDF
RSA Monthly Online Fraud Report -- February 2014
 
PPTX
Forex graphs
PPTX
Make your presentations stick (2): Magnficient Metaphors
PPT
Gedeelddoor pime
PDF
Broadview Mortgage Long Beach - California Extra Credit Teacher Program
PPTX
New world names lesson
PPS
Abortion
Mit2 092 f09_lec23
Discover Great Reasons to move to ConfigMgr 2012 SP1
Lessico civico. Teorie e pratiche della cittadinanza
Japan russio japanese war
Becerrajavier a1
elasticity 2014
How Does Long-term Care Insurance Work?
Day 3 mon world
What's new for Lync 2013 Clients & Devices
Wed thurs reform
TechBook: IMS on z/OS Using EMC Symmetrix Storage Systems
 
Mongol oron neelttei hicheel
Protectora d'animals_Xènia, Malina i Gemma
RSA Monthly Online Fraud Report -- February 2014
 
Forex graphs
Make your presentations stick (2): Magnficient Metaphors
Gedeelddoor pime
Broadview Mortgage Long Beach - California Extra Credit Teacher Program
New world names lesson
Abortion
Ad

Similar to White paper : Introduction to Fibre Channel over Ethernet (FCoE) - A Detailed Review (20)

PDF
Unified Fabric: Data Centre Bridging and FCoE Implementation
PPT
FCoE Origins and Status for Ethernet Technology Summit
PPTX
SAN Virtuosity Series: Network Convergence & Fibre Channel over Ethernet
PDF
City of Pforzheim case study
PPTX
Data Center: Cloud & Convergencia
PDF
PLNOG 13: Artur Pająk: Storage w sieciach Ethernet, czyli coś o iSCSI I FCoE
PDF
Instant Access to Fibre Channel for SANs 1st Edition Alan Frederic Benner ebo...
DOC
Noticias tel mayo 2011
PPTX
Snw Spring 09 F Co E Deployment In Virtualized Environments (031609) [9]
PDF
IndustryBrief - Unified Fabrics - Just Add Storage
PDF
SAN vs NAS technology summary
PDF
Fibre Channel for SANs 1st Edition Alan Frederic Benner
PDF
Fibre Channel for SANs 1st Edition Alan Frederic Benner
PDF
Unified Computing In Servers
PDF
Why Fibre Channel Fabrics to Support Flash Storage?
PDF
Analyst Perspective - Next Generation Storage Networking for Next Generation ...
PPTX
M4 f co-e_4.4.2
PPTX
Designing and deploying converged storage area networks final
PPT
Xtw01t6v0210 dcn
PDF
Sharath Pass 111111111111111111111111111111111111111111111111111111.pdf
Unified Fabric: Data Centre Bridging and FCoE Implementation
FCoE Origins and Status for Ethernet Technology Summit
SAN Virtuosity Series: Network Convergence & Fibre Channel over Ethernet
City of Pforzheim case study
Data Center: Cloud & Convergencia
PLNOG 13: Artur Pająk: Storage w sieciach Ethernet, czyli coś o iSCSI I FCoE
Instant Access to Fibre Channel for SANs 1st Edition Alan Frederic Benner ebo...
Noticias tel mayo 2011
Snw Spring 09 F Co E Deployment In Virtualized Environments (031609) [9]
IndustryBrief - Unified Fabrics - Just Add Storage
SAN vs NAS technology summary
Fibre Channel for SANs 1st Edition Alan Frederic Benner
Fibre Channel for SANs 1st Edition Alan Frederic Benner
Unified Computing In Servers
Why Fibre Channel Fabrics to Support Flash Storage?
Analyst Perspective - Next Generation Storage Networking for Next Generation ...
M4 f co-e_4.4.2
Designing and deploying converged storage area networks final
Xtw01t6v0210 dcn
Sharath Pass 111111111111111111111111111111111111111111111111111111.pdf

More from EMC (20)

PPTX
INDUSTRY-LEADING TECHNOLOGY FOR LONG TERM RETENTION OF BACKUPS IN THE CLOUD
 
PDF
Cloud Foundry Summit Berlin Keynote
 
PPTX
EMC GLOBAL DATA PROTECTION INDEX
 
PDF
Transforming Desktop Virtualization with Citrix XenDesktop and EMC XtremIO
 
PDF
Citrix ready-webinar-xtremio
 
PDF
EMC FORUM RESEARCH GLOBAL RESULTS - 10,451 RESPONSES ACROSS 33 COUNTRIES
 
PPTX
EMC with Mirantis Openstack
 
PPTX
Modern infrastructure for business data lake
 
PDF
Force Cyber Criminals to Shop Elsewhere
 
PDF
Pivotal : Moments in Container History
 
PDF
Data Lake Protection - A Technical Review
 
PDF
Mobile E-commerce: Friend or Foe
 
PDF
Virtualization Myths Infographic
 
PDF
Intelligence-Driven GRC for Security
 
PDF
The Trust Paradox: Access Management and Trust in an Insecure Age
 
PDF
EMC Technology Day - SRM University 2015
 
PDF
EMC Academic Summit 2015
 
PDF
Data Science and Big Data Analytics Book from EMC Education Services
 
PDF
Using EMC Symmetrix Storage in VMware vSphere Environments
 
PDF
Using EMC VNX storage with VMware vSphereTechBook
 
INDUSTRY-LEADING TECHNOLOGY FOR LONG TERM RETENTION OF BACKUPS IN THE CLOUD
 
Cloud Foundry Summit Berlin Keynote
 
EMC GLOBAL DATA PROTECTION INDEX
 
Transforming Desktop Virtualization with Citrix XenDesktop and EMC XtremIO
 
Citrix ready-webinar-xtremio
 
EMC FORUM RESEARCH GLOBAL RESULTS - 10,451 RESPONSES ACROSS 33 COUNTRIES
 
EMC with Mirantis Openstack
 
Modern infrastructure for business data lake
 
Force Cyber Criminals to Shop Elsewhere
 
Pivotal : Moments in Container History
 
Data Lake Protection - A Technical Review
 
Mobile E-commerce: Friend or Foe
 
Virtualization Myths Infographic
 
Intelligence-Driven GRC for Security
 
The Trust Paradox: Access Management and Trust in an Insecure Age
 
EMC Technology Day - SRM University 2015
 
EMC Academic Summit 2015
 
Data Science and Big Data Analytics Book from EMC Education Services
 
Using EMC Symmetrix Storage in VMware vSphere Environments
 
Using EMC VNX storage with VMware vSphereTechBook
 

Recently uploaded (20)

PPTX
MYSQL Presentation for SQL database connectivity
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PPTX
Group 1 Presentation -Planning and Decision Making .pptx
PDF
Encapsulation_ Review paper, used for researhc scholars
PDF
Approach and Philosophy of On baking technology
PPTX
Programs and apps: productivity, graphics, security and other tools
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Accuracy of neural networks in brain wave diagnosis of schizophrenia
PPTX
A Presentation on Artificial Intelligence
PPTX
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
PPTX
1. Introduction to Computer Programming.pptx
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
Getting Started with Data Integration: FME Form 101
PPTX
Big Data Technologies - Introduction.pptx
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
MYSQL Presentation for SQL database connectivity
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Building Integrated photovoltaic BIPV_UPV.pdf
Group 1 Presentation -Planning and Decision Making .pptx
Encapsulation_ Review paper, used for researhc scholars
Approach and Philosophy of On baking technology
Programs and apps: productivity, graphics, security and other tools
Per capita expenditure prediction using model stacking based on satellite ima...
Accuracy of neural networks in brain wave diagnosis of schizophrenia
A Presentation on Artificial Intelligence
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
1. Introduction to Computer Programming.pptx
Dropbox Q2 2025 Financial Results & Investor Presentation
Advanced methodologies resolving dimensionality complications for autism neur...
The Rise and Fall of 3GPP – Time for a Sabbatical?
Getting Started with Data Integration: FME Form 101
Big Data Technologies - Introduction.pptx
Mobile App Security Testing_ A Comprehensive Guide.pdf

White paper : Introduction to Fibre Channel over Ethernet (FCoE) - A Detailed Review

  • 1. Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review Abstract Fibre Channel over Ethernet (FCoE) is a new storage networking option, based on industry standards, which is now being deployed in real world environments. This white paper provides an overview of FCoE, describes the hardware and software components that make up the ecosystem, and explains how the technology is expected to continue to mature over the next few years. April 2011
  • 2. Copyright © 2008, 2009, 2010, 2011 EMC Corporation. All rights reserved. EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. THE INFORMATION IN THIS PUBLICATION IS PROVIDED ―AS IS.‖ EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Use, copying, and distribution of any EMC software described in this publication requires an applicable software license. For the most up-to-date listing of EMC product names, see EMC Corporation Trademarks on EMC.com VMware and ESX are registered trademarks or trademarks of VMware, Inc. in the United States and/or other jurisdictions. All other trademarks used herein are the property of their respective owners. Part Number h5916.3 Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 2
  • 3. Table of Contents Executive summary ............................................................................................ 4 Introduction ......................................................................................................... 5 Audience ...................................................................................................................................... 6 Creation of an FCoE infrastructure ................................................................... 6 FCoE has FC inside ............................................................................................ 6 FCoE is FC over ‘Lossless Ethernet’ ................................................................ 7 Road to a converged fabric ................................................................................ 8 Conclusion ........................................................................................................ 10 Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 3
  • 4. Executive summary The question for most data center managers is not should they network their storage, but which of the many options available should they use and for which applications can they justify the expense. Using networked storage reduces not only capital expenses through higher utilization of resources and the ability to create highly available configurations, but also operational expenses by allowing centralized management and simplified backup and replication. While most companies are using networked storage, overall less than 20 percent1 of all data center servers are attached to networked storage. Virtualized servers have over an 80 percent2 network attach rate since many advanced features such as mobility and high availability cannot be utilized if these servers are not using networked storage. It is expected that as the overall trend of server virtualization increases that the percentage of servers attached to networked storage in the data center will increase. One of the reasons why more servers do not utilize networked storage is that the storage network (Fibre Channel) is separate from the Ethernet (TCP/IP) network (see Figure 1). Fibre Channel is the predominate choice for networked storage in the data center because it allows for the creation of a highly scalable – hundreds or thousands of nodes – and reliable solution that can be centrally managed. For networking, today’s servers typically have multiple 1 Gb Ethernet network interface cards (NICs). The number of NICs in a system will vary depending on the applications on the machine, and management and security requirements, but the number can be four, six, eight, or more NICs in a single machine. This high number of server adapters, cables, and separate networking and storage fabrics adds to complexity and cost. As a result, many customers would like to be able to have all applications run over a single converged network. This goal of I/O consolidation on a unified network for all traffic leads to a savings in infrastructure (cabling, reduced sparing, and so on) and a simplified management environment. Overall data center trends require solutions that take up less space, draw less power, and require less cooling; suppliers are reaching a breaking point where supporting multiple separate networks will not allow them to meet these requirements. Figure 1. Existing data center with separate Ethernet and FC networks 1 IDC, Worldwide Fibre Channel Host Bus Adapter 2007–2011 Forecast and 2006 Vendor Shares and Worldwide and Regional Server 2007-2011 Forecast 2 Enterprise Storage Group, Server Virtualization: The Impact on Storage, November 7, 2007 Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 4
  • 5. Ethernet has been given top consideration as a potential solution for I/O consolidation due to the large install base and broad general understanding of the technology by all IT departments. While some storage applications can run comfortably with 1 Gb bandwidth widely available with Ethernet, many data center solutions require the higher bandwidth solutions that Fibre Channel can supply today with 2 Gb, 4 Gb, 8 Gb, and 16 Gb options. With the maturation of 10 Gb Ethernet, however, there is an inflection point where we have the opportunity to consolidate the many existing connections in a server to a pair of 10 Gb Ethernet links. Blade servers and server virtualization solutions, as well as newer server bus architectures like PCI Express 2.0, will allow infrastructure to utilize the bandwidth of 10 Gb adapters. The challenge is in unifying storage traffic and network traffic onto the same link. Fibre Channel over Ethernet (FCoE) will allow an Ethernet-based SAN to be introduced into the FC-based data center without breaking existing administration tools or workflows. This is done by allowing Fibre Channel traffic to run over an Ethernet infrastructure. Servers and storage see FCoE as FC (Figure 2) since FCoE is simply encapsulating FC over the Ethernet portions of the connectivity, while to the server applications as well as FC SAN and FC storage, the traffic is still FC. FCoE can be easily added into existing FC environments in contrast to previous attempts at consolidation with iSCSI and InfiniBand, which required new drivers and a rip-and- replace for customers that are using Fibre Channel. The security and management best practices of having a single application per NIC will need to be revisited as part of 10 Gb Ethernet adoption. Ethernet Network FC Network Driver Driver Converged Network Adapter Encapsulate FC storage Lossless Ethernet Links FC network FCoE Switch Ethernet De-Encapsulate FC Figure 2. FCoE environment with converged adapters at the server attached to a converged switch Introduction Today’s networks use different protocols to send information between devices. Traditional Ethernet is a family of frame-based computer networking technologies for local area networks (LANs), whereas Fibre Channel is used for storage area networking (SANs). Fibre Channel over Ethernet, or FCoE, is a new storage networking protocol that supports Fibre Channel natively over Ethernet. FCoE encapsulates Fibre Channel frames into Ethernet frames, allowing them to run alongside traditional Internet Protocol (IP) traffic. This white paper provides an overview of FCoE, describes the hardware and software components that make up the new ecosystem, and explains how the technology is expected to mature over the next few years. Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 5
  • 6. Audience This white paper is intended for network and storage administrators who want to learn more about FCoE and its benefits. Creation of an FCoE infrastructure FCoE requires the deployment of three new components: a Converged Network Adapter (CNA), Lossless Ethernet Links, and an FCoE switch. The CNA provides the functions of both a standard NIC and a FC HBA in a single adapter in the server. There are two types of CNAs available: a ―hardware‖-based solution — where the lower-level FC (exchange and sequence management) and FCoE functions are done in hardware — or a ―software‖ solution — where some of those functions are performed in software. When obtaining a hardware solution from a vendor that manufacturers FC HBAs, the drivers will be the same for both FC and FCoE adapters; this provides the least disruption and easiest integration into existing environments. The software solutions require the development of new FC drivers that will be run in software; one of the efforts can be found at Open-FCoE.org. The FCoE switch is the network device that will connect to existing LAN and SAN environments. The T11 standards FC-BB-5 group (http://www.t11.org/fcoe) created the FCoE protocol, which enables the creation of CNAs and FCoE switches. The link that FCoE traffic uses must provide the same level of lossless behavior that can be found with Fibre Channel environments today. The Data Center Bridging group (part of the IEEE 802.1 standards, see http://www.ieee802.org/1/pages/dcbridges.html) is investigating requirements to create an Ethernet environment that can support storage and IPC traffic. The enhancements to Ethernet will be discussed later, but new functionality is required in the devices and 10 Gb Ethernet is going to be the entry- level solution. Customers have started to introduce 10 Gb Ethernet in their data centers, so the FCoE effort is poised to help mold the technology and catch the incoming wave of adoption. FCoE has FC inside In creating the FCoE protocol, the goal is summed up in the name — take FC and put it over an Ethernet infrastructure. To do this, the Fibre Channel frames are encapsulated — not translated or bridged — in an Ethernet frame (Figure 3). The mapping is 1:1, which means that there is no segmenting of FC frames nor are multiple FC frames put in a single Ethernet frame. It is a requirement that all devices for FCoE (adapters and switches along the path) support mini-jumbo frames to allow the largest FC frame to be supported without segmentation. Flow control of FCoE is controlled by Priority Flow Control mechanisms rather than by the buffer credits of FC. The naming conventions for FCoE port types are very similar to FC: N_Ports (HBAs and storage) are VN_Ports (CNAs or FCoE storage). F_Ports (fabric ports) are VF_Ports (FCoE switch ports that attach to VN_Ports). E_Ports (switch to switch) are VE_Ports (ports between two FCoE Ethernet switches). Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 6
  • 7. Figure 3. FCoE frame format In Fibre Channel, security is typically not high on the list of discussion points. In the early days of FC, the small-scale physical security and optical cabling were usually sufficient for customers. The Fibre Channel Security Protocols standard (FC-SP) was approved in 2004 and addresses how to protect against security breaches. Ethernet has its own well-established security practices, so the standards were only concerned with specific threats that are not covered in a typical Ethernet environment. An issue of concern was that while FC links are always private (two ends), FCoE links can be shared (three-plus ends) due to the nature of Layer 2 Ethernet. Access Control Lists (ACLs) will be used to specify on a port basis what operations are allowed to be performed, similar to the protection provided by a firewall. A feature called ―Dynamic ACLs‖ is supported by the FCoE switches to reduce the administrative burden of manually configuring ACLs. FC-BB-5, the T11 standard for FCoE was ratified in June 2009. FCoE is FC over ‘Lossless Ethernet’ The Ethernet infrastructure, over which FCoE will travel, must be of a lossless nature. Since the FCoE protocol does not contain TCP, any lost packets would require recovery at the SCSI layer. In a typical environment, based on existing Ethernet, this would happen much too often to be usable for storage environments. Fortunately, the Ethernet community has been looking at the issue of creating a lossless environment for a number of years. The set of features required to support lossless behavior is referred to as Data Center Bridging (DCB) and is being worked on by the IEEE DCB working group. The first set of features is link level protocols and by definition only covers the link between two devices (that is, either an adapter and switch port or two switch ports); they are Priority Flow Control (PFC) and Enhanced Transmission Selection (ETS). Even before FCoE was conceived, there was a PAUSE function (Link Pause, IEEE 802.3, Annex 31B) that would allow for the creation of a lossless environment by stopping all traffic on a port when a full queue condition was achieved. The problem with classical PAUSE is that you cannot specify which traffic should be lossless, and as a result there is no mechanism to deal with the congestion created on all classes of traffic when a PAUSE is issued. Priority Flow Control (IEEE 802.1Qbb) creates a new PAUSE function that can halt traffic according to the priority tag while allowing traffic at other priority levels to continue. Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 7
  • 8. Administrators use the eight lanes defined in IEEE 802.1p to create virtual lossless lanes for traffic classes like storage (that require lossless behavior) and lossy lanes for other classes. In order to make sure that one traffic stream does not take too much of the overall bandwidth, the addition of Enhanced Transmission Selection (IEEE 802.1Qaz) is important for creating an environment where storage and other traffic can share the same link. A common management framework can be created for bandwidth management. High-priority traffic like storage can be prioritized and allocated bandwidth while still allowing other traffic classes to utilize the bandwidth when it is not being fully utilized. The Data Center Bridging Exchange Protocol (DCBX) is responsible for the configuration of link parameters for Data Center Bridging functions. It determines which devices support the enhanced functionalities that create the ―DCB‖ cloud where FCoE traffic can safely travel. It also allows for transparent pass-through for non-DCB traffic classes. These link level enhancements start to create a safe environment for FCoE traffic. PFC, ETS, and DCBX have Virtualization and FCoE all been defined and therefore solutions that have these Currently, server virtualization functions shipping today should be able to support FCoE. environments (including VMware and Microsoft Hyper-V) can choose among As the configurations expand, there are additional Ethernet FC, iSCSI, and NAS for networked enhancements beyond link level that will increase storage. The hypervisor has a storage performance. Congestion Notification (IEEE 802.1Qau) is driver stack that presents FC (from an a development to manage congestion beyond the link level. HBA) and iSCSI (from a NIC) traffic to This is desirable when a link is reaching a PAUSE individual virtual machines (VMs) or condition and the source of the traffic can be alerted and guests as storage. There is also a virtual throttled appropriately, reducing congestion out of the switch in the hypervisor that sends network core. While not necessary to support FCoE, an traditional LAN traffic to the VM/guest as additional enhancement to optimize Ethernet for storage network traffic. FCoE solutions using a environments is Layer 2 multipathing. Ethernet solutions CNA will function equivalently as today use Spanning Tree Protocol (STP), which only existing solutions – the CNA will convert supports a single active path between any two network FCoE traffic to FC packets in the nodes. The IETF standards group has created TRILL hardware, so the hypervisor and VM/guest (TRansparent Interconnection of Lots of Links), which will still work as if physical FC HBAs and allows for increased bandwidth by allowing and NICs were installed. Software FCoE aggregating multiple network paths. solutions with a standard NIC will require additional developments before they can be used. Currently hypervisors will not Road to a converged fabric identify FCoE traffic that is still Many customers are already running their storage traffic encapsulated and virtual switches over Ethernet by utilizing iSCSI or NAS; FCoE is not (including the Cisco Nexus 1000V in designed to replace these solutions. For customers that VMware® ESX®) do not have Lossless have an existing infrastructure and knowledge base of FC, Ethernet functionality, so sending traffic FCoE provides a path toward reaching a converged fabric. directly to the VM/guest would not be FCoE, iSCSI, and NAS will all leverage 10 Gigabit reliable. FCoE will expand the available Ethernet. In general, iSCSI environments tend to be small storage networking solutions options for configurations with under 100 servers, while most FC the high-growth server virtualization customers are scaling into hundreds or thousands of nodes; market. Moving to a converged fabric FCoE can be plugged in to these existing environments. will simplify mobility by reducing the 10 Gigabit Ethernet will provide the common number of networks that need to be infrastructure that can support FCoE, iSCSI, and NAS moved. It also allows a standard (Figure 4), all of which can be used in virtualization configuration to be built, allowing for environments, and customers will choose based upon their flexible storage deployment. requirements and skill set. Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 8
  • 9. 10 Gigabit Ethernet Fibre Channel common management common infrastructure virtualization Figure 4. 10 Gigabit Ethernet common infrastructure for storage All new technologies take time to develop a robust ecosystem, and FCoE is no exception. Since FCoE builds on the existing infrastructure, it can be rolled out in a phased manner. Today FCoE can support convergence at the server, in which the separate network (NIC) and storage (HBA) interfaces on the server are replaced by the CNA that is directly attached to the FCoE switch. A new rack of servers can deploy these technologies, while the existing storage and network environments remain unchanged. FCoE can also support convergence at the storage port and on switch-to-switch links, allowing for an end-to-end (FCoE multi-hop) configuration to be created. Recently, FCoE solutions started to expand beyond a single switch at the access layer into multi-hop configurations. While the industry continues to work on the technical issues of expanding FCoE solutions, customers should review the structure of their storage and networking groups to make sure that there are not any ―political‖ issues over who owns the purchase, installation, and maintenance of the CNA and FCoE switch components in a converged network environment. Some of the enhancements to Ethernet will be new to network administrators who will need to learn these functions and will require coordination with storage administrators. The overlapping domains may compel cultural adjustments, as storage networks will no longer be dedicated and network configurations can no longer be reconfigured at-will. Network and storage administrators will each have their own management interfaces to the environments, keeping tasks separate rather than converged. As director-class products that support FCoE are released into the market, building FCoE multi-hop configurations that include native FCoE storage begins to make sense. No discussion of FCoE is complete without addressing the cabling infrastructure supporting the solution. When customers deploy a physical cabling infrastructure, they are looking for solutions that can be used for five to ten years. For 1 Gigabit Ethernet, the primary options for cabling are copper (1000Base-T with RJ- 45 connectors) and optical (same physical cabling as FC); copper dominates this market with billions of ports installed while optical has historically been 1 percent of Ethernet ports3. A standard (10GBase-T) for using existing copper cabling (either CAT 6 or CAT 6a with RJ-45 connectors) is available but not expected to start shipping for FCoE solutions until 2012. A new copper option known as Twinax has become available for FCoE solutions of 10 Gigabit Ethernet. The option is based on the SFF-8431 standard and uses the SFP+ interface for a copper connection that is low cost and low power. The Twinax solution is limited to short distances, supporting between 1-10 meters, which is sufficient for server to top- 3 IEEE Installed Cabling market data (http://ieee802.org/3/10GBT/public/jan03/flatman_1_0103.pdf) Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 9
  • 10. of-rack or end-of-row switch environments. Standard multimode optical cabling will be used for environments that require longer distances such as from the rack to the core. Conclusion FCoE is a natural extension of FC networks that will allow data centers to leverage converged I/O and extend the operational efficiencies of networked storage to a new wave of servers. Server, storage, and networking vendors have all invested heavily to create converged fabric technologies. There is a robust FCoE ecosystem that will continue to expand and mature over time. Introduction to Fibre Channel over Ethernet (FCoE) A Detailed Review 10