Issue 2/2014

Transcription

Issue 2/2014
The communications technology journal since 1924
Communications as a cloud service: a
new take on telecoms 4
Capillary networks – a smart way to get
things connected 12
Trusted computing for infrastructure 20
Wireless backhaul in future
heterogeneous networks 28
Connecting the dots: small cells shape
up for high-performance indoor radio 38
Architecture evolution for automation
and network programmability 46
2/2014
Editorial
Deeper into the
Networked Society
Earlier this year, Ericsson launched
its new vision: a Networked Society
where every person and every industry is empowered to reach their full
potential. Technology leadership is
about realizing this vision. It’s about
developing connectivity technology
to make it an integral part of our daily
lives, whether we’re at work, at school,
at home, outside, on the way somewhere or taking part in some event.
The aims of each individual or enterprise vary widely; they want coverage,
capacity, reliability, availability and
resilience with an appropriate level of
security. The one-size-fits-all network
model no longer applies; network characteristics need to be tailored to users’
specific needs. With cloud technologies, SDN and NFV as a foundation, the
technological developments we are
working on – in the move toward 5G –
are based on providing connectivity to
suit every different use case.
The traditional way of building services and applications by packaging
functionality and data and inherently
assuring security has worked well for
services and applications made and
delivered by just one vendor – some
even benefiting from bundling with
hardware. But this approach doesn’t
lend itself to the creation of innovative
solutions that provide benefit. Nor does
it fit with reusability, fast time to market, and the use of generic hardware.
Instead, applications are being
mashed together from lots of other
internet services. However, the freedom to innovate that this approach
offers leads to security issues, which
is one of our industry’s greatest challenges. And as web services and programmable routing technology are
deployed on platforms that exploit virtualization, assuring security becomes
trickier still.
In the face of such challenges,
trusted computing helps us to meet
the evolving security requirements of
E R I C S S O N R E V I E W • 2/2014
users, businesses, regulators and infrastructure owners.
The developments that I would
like to highlight relate to handling
expected growth in traffic volumes, capacity and machine-type
communication.
Building heterogeneous networks
is an effective way of expanding
networks to handle traffic growth.
However, the additional small cells
included in these networks need to be
provided with flexible and cost-efficient
backhaul. Our research shows that nonline-of-sight backhaul in licensed spectrum is a future-proof technology in
this area.
When it comes to capacity, one of
the significant challenges is providing radio capacity indoors. About 70
percent of all traffic is generated
indoors, and our research has resulted
in a novel small cell solution with a
flexible radio architecture. We wanted
to address the issue of indoor capacity
from an ecosystem point of view, with
an emphasis on cost control at every
phase. From installation to operation,
our aim was to create a special indoor
small cell that works well in large
buildings – a solution that would integrate smoothly with outdoor coverage.
Capillary networks offer a smart
way to connect the Internet of Things,
but they require some additional functionality. The use cases for machinetype communication vary greatly
from one application to the next, and
so rather than building systems with
a one-size-fits-all approach, capillary
networks will be designed to fit the
application.
All of these developments lead to
the establishment of a flexible network
architecture set to satisfy the demands
of every future use case. As always, I
hope you enjoy our insights.
About 50 percent of all
sites will be connected with
microwave in 2019.*
*Ericsson Mobility Report, June 2014
Ulf Ewaldsson
Chief Technology Officer
Head of Group Function
Technology at Ericsson
The communications technology journal since 1924
CONTENTS
2/2014
Communications as a cloud service: a
new take on telecoms 4
Capillary networks – a smart way to get
things connected 12
Trusted computing for infrastructure 20
Wireless backhaul in future
heterogeneous networks 28
Connecting the dots: small cells shape
up for high-performance indoor radio 38
Architecture evolution for automation
and network programmability 46
To bring you the best of Ericsson’s
research world, our employees have
been writing articles for Ericsson
Review – our communications
technology journal – since 1924.
Today, Ericsson Review articles have
a two-to-five year perspective and
our objective is to provide you with
up-to-date insights on how things are
shaping up for the Networked Society.
Address :
Ericsson
SE-164 83 Stockholm, Sweden
Phone: +46 8 719 00 00
Communications as a cloud service:
a new take on telecoms
4
Software as a service (SaaS) is a promising solution for overcoming the challenges of
implementing and managing new network technologies and services like voice over LTE (VoLTE).
The SaaS approach can provide substantial savings in terms of cost and lead-time, and create a
new source of revenue for service providers.
This article was originally published on July 22, 2014.
Capillary networks – a smart way to
get things connected
12
A capillary network is a local network that uses short-range radio-access technologies to provide
local connectivity to things and devices. By leveraging the key capabilities of cellular networks
– ubiquity, integrated security, network management and advanced backhaul connectivity –
capillary networks will become a key enabler of the Networked Society..
This article was originally published on September 9, 2014.
20
Publishing:
Ericsson Review articles and
additional material are published
on www.ericsson.com/review. Use
the RSS feed to stay informed of the
latest updates.
Articles are also
available on the Ericsson
Technology Insights app
for Android and Apple
devices. The link for your device is on
the Ericsson Review website:www.
ericsson.com/review. If you are
viewing this digitally, you can:
download from Google Play or
download from the App Store
Publisher: Ulf Ewaldsson
Editorial board:
Håkan Andersson, Hans Antvik,
Ulrika Bergström, Joakim Cerwall,
Stefan Dahlfort, Deirdre P. Doyle,
Dan Fahrman, Anita Frisell,
Jonas Högberg, Patrik Jestin,
Magnus Karlsson, Cenk Kirbas,
Sara Kullman, Börje Lundwall,
Hans Mickelsson,Patrik Regårdh,
Patrik Roséen and Gunnar Thrysin
Editor:
Deirdre P. Doyle
deirdre.doyle@jgcommunication.se
Contributors:
John Ambrose, Paul Eade,
Nathan Hegedus, Ian Nicholson,
Ken Neptune and
Birgitte van den Muyzenberg
Art director and layout:
Carola Pilarz
Illustrations:
Claes-Göran Andersson
Printer: Edita Bobergs, Stockholm
2/20143
Trusted computing for infrastructure
Modern internet services rely on web and cloud technology, and as such they are no longer
independent packages with in-built security, but are constructed through the combination
and reuse of other services distributed across the web. While the ability to build applications
in this way results in highly innovative services, it creates new issues in terms of security.
Trusted computing aims to provide a way to meet the evolving security requirements of users,
businesses, regulators and infrastructure owners.
This article was originally published on October 24, 2014.
28
Wireless backhaul in future heterogeneous networks
Heterogeneous networks are an effective way of expanding networks to handle traffic growth.
However, the additional small cells included in heterogeneous networks need to be provided with
backhaul – in a way that is flexible and cost-efficient. Our research shows that non-line-of-sight
(NLOS) backhaul in licensed spectrum up to 30GHz is a future-proof technology for managing
high volumes of traffic in heterogeneous networks.
This article was originally published on November 14, 2014.
Connecting the dots: small cells shape up for highperformance indoor radio
38
How do you design a small radio to fit the interiors of large spaces, yet powerful enough to meet
future requirements for indoor radio capacity? This was the question we asked ourselves when
we began to develop a solution to provide high-capacity radio for indoor environments.
This article was originally published on December 19, 2014.
Architecture evolution for automation
and network programmability
46
Automation and network programmability are key concepts in the evolution of telecom networks.
Architecture designed with high degrees of automation and network programmability can rapidly
adapt to emerging requirements, and as such improve operational efficiency and time to market
for new services.
This article was originally published on November 28, 2014.
ISSN: 0014-0171
Volume: 91, 2014
E R I C S S O N R E V I E W • 2/2014
Proof of concept for VoLTE as a service
4
Communications as a cloud
service: a new take on telecoms
Modern mobile networks are complex systems built with an increasingly broad variety of technologies
to serve a wide base of devices that provide an ever-greater range of services. These developments
create interesting business opportunities for operators. But they also bring challenges, as new
technologies and new expectations need to be managed with the same staff and budget.
BA RT J E L L E M A A N D M A RC VORW E R K
Software as a service (SaaS)
is a promising solution for
overcoming the challenges of
implementing and managing new
network technologies. The SaaS
approach can provide substantial
savings in terms of cost and lead
time, and create a new source of
revenue for those adopting the
role of service provider.
This article shares some of the technical
and economical insights and know-how
gained from a proof of concept study
conducted at Ericsson to explore the
implementation of VoLTE as a service.
Why a new take on telecoms?
Today’s networks support several technology generations, from 2G to 4G, and
as research for 5G is well underway,
the next generation is on the commercial horizon. The types of devices connected to networks vary from feature
phones to smartphones and tablets to
the billions of new connected devices
that are emerging to support applications like smart homes and connected
vehicles. In short, this is a complex ecosystem based on constant development,
which can be difficult to predict and
consequently challenging to plan for
and budget.
The introduction of 4G LTE networks, for example, brought with it a
major overhaul of voice services in core
networks – in the move from circuitswitched to IMS. For many, especially
niche operators, this type of technology
upgrade threatens to stretch organizational capabilities to the limit, even to
the point where business profitability
is at stake.
To counter this challenge, many operators have turned to Network Functions
Virtualization (NFV). By placing core
networks in large concentrated data
centers, NFV is a way to rationalize and
simplify operations as well as speeding up innovation cycles1. The addition
BOX A Terms and abbreviations
ARPU
average revenue per user
CRM
customer relationship management
CSCF
Call Session Control Function
HSS
Home Subscriber Server
IMS
IP Multimedia Subystem
LI
Lawful Interception
MRFP
Media Resource Function Processor
MSC
mobile switching center
MTAS
Multimedia Telephony Application
Server
MVNO
mobile virtual network operator
NFV
Network Functions Virtualization
E R I C S S O N R E V I E W • 2/2014
NPV
O&M
OSS
OVF
P-CSCF
SaaS
SBG
SLA
SRVCC
TCO
VLAN
VM
VoLTE
net present value
operations and maintenance
operations support systems
Open Virtualization Format
proxy call session control function
software as a service
Session Border Gateway
Service Level Agreement
single radio voice call continuity
total cost of ownership
virtual local area network
virtual machine
voice over LTE
of multi-tenancy capabilities to NFV
makes this approach particularly interesting for global operators, who have a
presence in several countries and manage a range of networks through various
operating companies.
Apart from addressing the strain on
internal resources, NFV opens up the
opportunity for operators to provide services, like VoLTE, to other communication service providers. By deploying
the necessary IMS network functions
for services in a central virtualized data
center, and by adopting a SaaS model,
operators can unlock the potential of
their infrastructure beyond their own
portfolios. Virtualized services can then
be offered to smaller second and third
tier affiliates or MVNOs at a lower cost,
with reduced risk, and within a shorter
time frame than is normally associated with the introduction of new services using traditional telecom business
models.
The SaaS business model allows
an operator’s partners to circumvent
lengthy hardware procurement cycles.
This way, the burden of costs and complexities associated with owning a
completely new and technologically
advanced communications system can
be removed. Simply by signing up as a
tenant to the existing facilities of a host
operator’s data center, partners will
be able to provide services quickly and
cost-efficiently.
Once in place, NFV provides a flexible
telecom-grade platform on which a variety of communication services can be
offered to people and organizations, in
a low-cost, low-impact fashion. Services
can be quickly and easily trialed,
launched, scaled up or down and decommissioned in line with market demand,
5
presenting an operator-branded and
guaranteed alternative to the many
third-party over-the-top solutions that
operate in both the consumer and enterprise communication space.
Concept – heading for the clouds
Today, the purchasing process for a new
IMS system can take several months
from order placement to an operational system. Once an order is placed,
the network system vendor initiates
the production process for the node.
On completion, the node is then integrated and packaged together with the
necessary software elements, tested,
shipped, installed at the designated central office site, integrated into the network, tested again, accepted and finally
put into operation. Once the system is
functional the operator is responsible
for operations and maintenance (O&M),
often with the support of the vendor.
With a SaaS deployment, operators
can purchase a virtualized IMS network
slice that is custom-initialized for them
in a large data center. Network slices can
be tied into existing radio and packet
core networks over a remote link – as
Figure 1 illustrates.
Working in this way, operators will
no longer need to purchase, install or
own any hardware, or invest in training staff on a new system. The SaaS
approach removes the need to manage
software licenses, and reduces system
integration from a complete IMS solution to just the points of interconnect
with the access network. Ownership
and operational details are instead
taken care of by the service provider and
operators will pay as they go using simple, predictable price models, such as a
flat service fee per subscriber. The benefits: no large upfront investments, limited technical and business risks, and
much shorter time to revenue.
VoLTE as a service
In 2013, Ericsson’s R&D and IT divisions
carried out a joint project to develop a
proof of concept implementation for
VoLTE as a service. The objective was
to gain an understanding of the technical and economic implications of
offering a complex communications
solution like VoLTE as a service. For telecom applications, SaaS is a relatively
new business model that needs to
FIGURE 1 The SaaS concept
IMS
EPC
LTE RAN
Traditional node deployment
Software as a service
EPC
IMS
LTE RAN
FIGURE 2 VoLTE as a service – architecture
Cloud-based
multi-tenant
VoLTE system
Tenant Y
Tenant X
EMA
MM
MSP
HSS
PGM
MTAS
SCC-AS
CSCF
BGCF
MRFP
SBG
P-CSCF
DNS/
ENUM
Tenant X
BSS
MSC-S
MGCF
SMS-C
CRM
MGw
BGF
EPC
LTE
Tenant Y
BSS
MSC-S
MGCF
SMS-C
CRM
MGw
BGF
EPC
LTE
E R I C S S O N R E V I E W • 2/2014
Proof of concept for VoLTE as a service
6
FIGURE 3 Network Functions Virtualization – portfolio migration
High
Media distribution network
Control plane elements, CSCF, MSC
Gateways and appliances
Hosted
managed
services
Distributed cloud
Value
OSS, BSS
Home appliances
Edge router
EMS
Real time OSS/BSS
Home networking
Core
router
Radio
access
Low
Fixed
access
High
Low
Risk
(Technology maturity, performance requirements)
FIGURE 4 IP design
Central
storage
Tenant 2
network
E R I C S S O N R E V I E W • 2/2014
CSCF
cluster
HSS
cluster
MTAS
cluster
vRtr
vRtr
vRtr
PGM
IDNS
MFRP
Storage
DC
switch
/FW
O&M
Signaling
Media
Tenant 1
Core
Access
SBG
Access
Tenant 1
network
O&M
Tenant 2
NOC
Tenant 1 IMS network
CSCF
cluster
HSS
cluster
MTAS
cluster
vRtr
vRtr
vRtr
PGM
iDNS
MFRP
O&M
Signaling
Media
Tenant 2 IMS network
take into consideration the tough
requirements of the underlying cloud
infrastructure.
From the start of the project, it was
clear that turning VoLTE as a service
into a viable business proposition, with
competitive price levels and sound margins, would require the onboarding and
serving of new tenants to be simple, efficient and easily repeated.
Through virtualization techniques,
the hosting service provider can deploy
multiple VoLTE systems on the same
shared data center hardware, while
still guaranteeing each tenant their
own dedicated, logically separated virtual network. Such a multi-tenant cloud
infrastructure makes it possible for service providers not only to share hardware among tenants, but also O&M and
engineering staff. The resulting economy of scale is much more significant
than any individual small-scale installation could achieve.
To improve repeatability, a high
degree of business process automation
(auto-deployment and auto-scaling)
reduces the time and effort needed to
operate services, which in turn reduces
costs. And to ensure that customers get
what they pay for, the provision of relevant network statistics is essential for
billing and to provide proof of Service
Level Agreement (SLA) conformance.
A blueprint for the architecture
So how is this done? As shown in
Figure 2, the operator’s radio and
packet core networks as well as their
legacy circuit-switched network are
connected to a remote virtualized IMS
network within a cloud data center over
standardized interfaces for signaling,
O&M and media.
As illustrated in Figure 3, next generation systems will normally be fully
implemented as software without
any strong hardware dependencies.
Consequently, IMS server-type network
functions like CSCF and MTAS are natural candidates for cloud placement.
To optimize use of bandwidth, most
media handling will most likely continue to take place in the tenant network, with the possible exception of the
MRFP. Certain network functions, such
as HSS, can be placed either in the cloud
or in the tenant network, depending on
operator preference or to comply with
7
local regulatory requirements with
respect to user databases.
To integrate with the operator’s various business support, customer care and
other IT systems, the virtualized IMS
network will provide billing and provisioning capabilities.
When another operator becomes a
tenant, a copy of the virtualized IMS
network can be instantiated in the data
center and the whole onboarding process simply repeated.
For commercial deployment, at least
two data center locations are needed to
provide geo-redundancy. Alternatively,
the tenant could operate a single nonredundant system in their own network
and rely on a secondary virtualized system as an overflow and failover mechanism – geo-redundancy as a service.
As an additional offering, service providers can include smaller regional satellite sites that host the IMS media plane
nodes. In such topologies, the satellite
centers can be used to house not just
media gateways but also network functions like Lawful Interception for IMS
(LI-IMS), an anchor MSC for SRVCC and
/or an SBG/P-CSCF. Providing mediaplane nodes in this way reduces the
impact of introducing IMS to an operator’s existing core network to practically
nothing. Taking a coverage area the
size of North America as an example,
approximately 24 regional sites would
be required to provide this service.
A significant change
By the end of December 2015, roaming
fees within the European Union will
no longer exist; rates for voice calls and
data transmission will be the same as
in the subscriber’s home market2. This
drastic change for consumers is likely
to stimulate traffic and motivate operators across Europe to centralize their
core network infrastructures – as physical location will no longer influence
billing rates.
Hardware
From a hardware perspective, data
centers will need to be equipped with
enough servers to host virtualized versions of the number of tenant IMS networks anticipated. In addition, high
capacity physical IP switches and central storage will be needed. As hardware
is completely decoupled from software
FIGURE 5 Operations and maintenance view
Work orders, tickets,
and change requests
DC back office
NOC front office
Tenant
SLA report,
invoicing
Tenant
provisioning
CM
PM
FM
NIM
TM
Network
management
SLA monitoring,
service metering
Subscriber
billing
Dashboard
Mediation
Subscriber
provisioning
Service
activation
CDRs
Node manager
Cloud manager, including:
• SLA resolution
• Auto-deployment
• Auto-scaling
BOX B  
Legend for
Figure 5
CM – configuration
management
DC – data center
FM – fault
management
NIM – network
inventory
management
NOC – network
operations center
PM – performance
management
TM – ticket
management
Subscriber data
vIMS
OS
Hypervisor
Hardware
through virtualization middleware,
service providers have the freedom to
select the x86-based hardware of their
choice, as long as it meets the set target
specifications in terms of performance,
bandwidth and memory of the virtualized network functions – including
some virtualization overhead.
Operations and maintenance
As shown in Figure 4, the IP plan needs
to be designed so that each tenant has
their own set of dedicated VLANs – at
least for O&M, signaling and media –
that are separated from all the other
tenants to avoid interference and maintain security.
For O&M, the service provider’s back
office can perform tasks such as configuration management, performance
management, fault management
and network inventory management
through a managed services portal.
This is similar to the way network management works in the service provider’s own IMS network. The front office
can process work orders and change
requests received from tenants, tickets
from field engineers, and take care of
invoicing and SLA reporting.
As shown in Figure 5, tenants will
be provided with O&M access rights to
their specific network for provisioning subscribers and retrieving detail
records for charging. This access will
connect to the tenant’s back-end IT systems like CRM and billing systems, and
a dashboard function will allow the tenant to view key performance statistics
for their network.
Northbound interfaces from the
different virtualized network functions are generally not affected by
virtualization.
The exact implementation of the IMS
network – its internal structure, what
software and which release is used – is
entirely at the discretion of the service
provider. In other words, the implementation is transparent and of no real
concern to the tenant. Their only concern lies with the behavior of the service, the agreed service level and the
interfaces exposed at the points of
interconnection.
Features – under the hood
Multi-tenancy
Modern server blades house multiple
processor cores on which virtual
E R I C S S O N R E V I E W • 2/2014
Proof of concept for VoLTE as a service
8
FIGURE 6 Multi-tenancy
HW
vRouter
VM
VM
VM
VM
(CSCF)
VM (MTAS) VM
VM
VM
VM
(HSS)
VM
vRouter
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
vRouter
VM
VM
VM
VM
VM
VM
VM
VM
VM
vRouter
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
Hyp
Hyp
Hyp
#1
#2
#3
VM
VM
VM
VM
VM
VM
VM
VM
VM
(DNS)
(PGM)
(MRFP)
VM
VM
VM
VM
VM
VM
Tenant Z Tenant Y Tenant X
vRouter
vRouter
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
Hyp
Hyp
Hyp
Hyp
Hyp
Hyp
Hyp
Hyp
Hyp
#4
#5
#6
#7
#8
#9
#10
#11
#12
vRouter
VM
VM
VM
vRouter
vRouter
12 blades
machines (VMs) can be placed.
Virtualized IMS network functions
like CSCF or HSS use a number of virtual machines for traffic processing, as
shown in Figure 6, which act much like
a physical node with several blades as
part of a cluster. As illustrated, these
virtual machines should be spread horizontally over multiple blades, so that
the failure of one blade will never bring
down an entire node.
The remaining cores can then be used
for other network functions or even
other tenants.
Auto-deployment
Onboarding a new tenant sets a deployment function into motion. As shown in
Figure 7, this function executes an IMS
network deployment sequence using a
cloud orchestration tool in combination
with scripts that parse the customerspecific environment settings. Any necessary adaptations are executed inside
the deployed VMs.
To save time during the onboarding
process, tenant VLANs can and should
be prepared ahead of time. Software
images for each virtualized network
function are built and uploaded (in
advance) to the cloud manager in, for
example, Open Virtualization Format
(OVF)3, and are kept in storage. From
there, the deployment function can
instantly clone network functions for
new tenants.
FIGURE 7 Auto-deployment
1. Clone from
template
2. Post-config
DNS/
ENUM
Deployment
function
CSCF
Cloud
manager
CSCF
vAPP
E R I C S S O N R E V I E W • 2/2014
CSCF
vRouter
HSS
HSS
HSS
vAPP
vRouter
EMA
MTAS
vRouter
MTAS
To connect them to their pre-assigned
VLANs, the virtual machines are linked
to the appropriate port groups and powered on. The deployment function loads
a data transcript onto the VMs to create an operational virtualized network
function and configures the application
interfaces, so that they form an integrated IMS network. All of this post-configuration work can be scripted; and any
data transcript common to all tenants
can be included in the software image.
Once all the network functions and
connections between them are established, the next step is to connect the virtual IMS network to the tenant’s access
network and IT systems before provisioning the first users. The high degree
of preparation and process automation, together with the use of hardware
capacity available in the data center, and
prestorage of software images, results in
drastically reduced installation times.
The complete software installation for
an IMS network can be fulfilled in just
a few hours, compared with the several
days it would normally take to set up a
traditional central office environment
with physical nodes. Time to revenue
from contract signing to commercial
launch could be reduced to a matter of
weeks, rather than months.
Auto-scaling
The ability to scale networks is a key
business enabler. In the proof of concept project, the Ericsson team developed a controller function that worked
in conjunction with the cloud manager
to determine when and where networks
need to be scaled. As shown in Figure 8,
the controller continuously monitors
the average processor load on each of
the virtualized network functions, by
reading the load figures from the guest
operating system. This approach has
proven to be more accurate than using
the measurements provided by the
hypervisor, as the hypervisor cannot,
for example, determine the priority and
necessity of currently executed tasks
from the outside.
When the load for a particular network function like CSCF exceeds its
set upper limit, which can happen for
example during traffic peaks, the controller requests the cloud manager
to scale out. The cloud manager powers up another CSCF virtual machine,
9
3. Join cluster
1. Measure
load
CSCF
Controller
Cloud
manager
CSCF
2. Power on VM
CSCF
vRouter
FIGURE 9 Example time series
Time: 13:00:44
Finished scaling out
Number of TPs: 5
CSCF-CBA-012 load
Number of traffic processors
PL 4
PL 5
PL 6
PL 7
PL 8
100
PL 3
Service-level monitoring
SLAs are highly varied in nature, covering different aspects of a service such as
customer ticket turn-around times and
other logistical matters. As far as technical content is concerned, SLAs between
service providers and tenants are best
kept simple and transparent. Many network statistics can be made available
for information purposes, which is fine,
but the list of contracted KPIs that carry
financial implications are best kept as
simple as possible (see Table 1).
In its simplest form, billing tenants for the use of VoLTE as a service
can be based on the actual number of
active users during a given time period
– assuming a certain maximum traffic volume. The volume can be defined
in terms of the maximum number of
simultaneous sessions (the current
licensing model) or by average voice
minutes per subscriber. As voice
FIGURE 8 Auto-scaling
51
52
50
49
38
off
80
Processor load (%)
which then joins the existing cluster
and rebalances the traffic. Similarly, a
node can be scaled in during periods of
low traffic.
The user interface for the controller
allows engineering staff in the data center to set upper and lower processor-load
thresholds for scaling in and out of network functions. Additional parameters,
such as the minimum and maximum
number of traffic processors, can also
be set so that a node has a guaranteed
minimum redundancy without monopolizing more than its fair share of available resources.
Figure 9 shows an example time
series taken from a test session carried
out during the proof of concept project.
The scaling mechanism for the CSCF
kicked in just before the 12:58 time
stamp, as the processor load exceeded
the set maximum (indicated by the red
line). Three minutes later, at approximately 13:01, the CSCF was running on
a cluster of five instead of the original
four traffic processors.
Depending on the existing traffic
load, it takes between five and 10 minutes to add capacity automatically (by
scaling a virtualized network function out by one traffic processor) to a
live node in a virtualized data center.
In contrast, adding a physical hardware
board to a live physical node on such a
time scale is unimaginable.
60
40
20
0
12:42
12:44
12:46
12:48
12:50
12:52
12:54
12:56
12:58
13:00
Time
%
Table 1: SLA reporting
Service Metering
Tenant gets billed per number of users +
premium for traffic coverage
Service Level Monitoring
Tenant gets credited in case of failure to meet SLA
Key performance indicators
Key performance indicators
System availability (%)
Number of users
IMS registration time (msec)
Traffic volume
(average session duration and/or
number of concurrent sessions
IMS registration success ratio (%)
VoLTE setup success ratio (%)
E R I C S S O N R E V I E W • 2/2014
Proof of concept for VoLTE as a service
10
FIGURE 10 Pricing model
opex
opex
opex
capex
3-year system TCO
(amoritized over 36 months)
Year 1
Year 2
minutes readily translate to the payment plans offered by most operators,
this model is probably preferable for the
majority of tenants. Similar consumption indicators aligned with operator-toconsumer price models can be created
for all other services.
While threshold limits are good for
SLAs and planning, service providers
are not likely to cut off traffic when
an agreed maximum for a tenant is
reached – as long as continued service
does not overload the system or infringe
on other tenants. However, a premium
may be charged.
To keep service level monitoring relatively straightforward, the proof of concept project created example reports for
system availability, registration time,
registration success rate and call establishment success rate. If any of these
Year 3
resources underperformed during a
billing period, the tenant would receive
credit on their next payment.
All of these counters and statistics
are already available in today’s typical
IMS products. By collecting, filtering
and combining them into a customized
business intelligence report, they can be
easily communicated and turned into
actionable data.
In a commercial setup, this data
would be fed from the OSS into a specialized SLA management tool, in which
KPI values are continuously compared
against predefined thresholds to detect
and record SLA violations. A number
of approach warning levels are usually
defined below the critical level, so that
O&M staff can be alerted and take appropriate actions before any impact on business is felt.
Table 2: TCO comparison – an example in USD thousands
system
capex
opex as a service
Hardware
2,400
1,000
Setup fee
Software
3,300
2,500
Service fee*
Systems Integration
1,600
Project
opex
450
24,098
900
Staff
4,300
Facilities
500
Utilities
Lab costs
capex
200
5,000
3 year TCO
21,700 3 year TCO
24,548
NPV
20,791 NPV
20,791
*36 months x 200,000 subscribers x USD 3.35
E R I C S S O N R E V I E W • 2/2014
Financials – where is the money?
In the traditional system-sales model,
total cost of ownership (TCO) is defined
as the initial purchase price including
related project costs, plus recurring running costs such as support agreements,
O&M staff, rent and power. In the SaaS
model, this will be replaced by a single
line item – service fees – under opex.
Unfortunately, estimating a reasonable price level for VoLTE as a service –
one that the tenant can afford and that
keeps the service provider in business –
is not a simple task.
One potential pricing model (shown
in Figure 10) is based on the traditional total cost of ownership for a threeyear period, amortized over 36 equal
monthly payments. Payback times of
less than three years tend to result in a
service that is too expensive for the tenant, and calculating over longer periods
tends to make the model unattractive
for service providers. Parameters like operator size and
running costs – rents, engineer salaries
and electricity – vary greatly from one
part of the world to another, and so the
economy of scale and benefit to operators in different markets will vary. In
conjunction with the proof of concept
project, a study aimed to estimate the
service price for VoLTE for a typical second or third tier operator with between
100,000 and one million subscribers.
The study estimated and adjusted for
net present value (NPV) and the required
initial capex and opex over three years
to own, deploy and run an IMS system
for VoLTE. The resulting estimation set
the fee for VoLTE as a service to be somewhere between USD 1 and USD 5 per
subscriber per month. An example of
the type of calculation used in the study
is given in Table 2 for a mock tenant
with 200,000 subscribers.
To match the price points for a service with the average cost per subscriber
incurred by operators at different ends
of the scale, some sort of tiered price
model is needed – a suggested model is
shown in Figure 11.
If the average revenue for voice services is assumed to be USD 40 per subscriber per month, a fee of USD 1-5 per
subscriber per month for VoLTE as a service is between 2.5 and 12.5 percent of
the corresponding ARPU it generates,
which is a fair business case.
11
Looking at the addressable market,
the number of subscribers connected to
second and third tier operators amounts
to 22 million in North America alone.
Evolution – beyond the horizon
As illustrated in Figure 12, rolling
out VoLTE might be the initial motivation for a second or third tier operator
to switch to the software as a service
model. Doing so would allow such operators to roll out VoLTE in the same time
frame (2014-2015) as their larger competitors – and secure their market share.
Subsequently, operators could
broaden the scope of their offerings to
include customized services for enterprises, the retail industry and many
other verticals. The SaaS platform could
be further utilized by opening it up to
internet-application and web developers to create a whole new range of converged services.
Second and third tier operators are
the most obvious first adopters of this
type of business model for voice – or
rather VoLTE. Once rooted, adoption is
likely to rise up the food chain. Many
operators, both big and small, have
opted for the managed service approach
for their voice networks, gaining efficiency and freeing up resources to focus
on customers and on improving operator brand value.
Operators already include unlimited
voice and unlimited text in their data
plans, rendering these services to the
level of a commodity, or a fundamental
product that cannot really be charged
for, but neither can they be taken out of
the service offering. And so software as
a service – the ultimate form of a managed service – is the most natural evolution path.
Conclusion
By following this route, service providers will be able to offer all managed
networks from the same platform and
housed under the same roof. For operators, the ability to outsource the responsibility for voice shifts price pressure
on to a third party who can provide the
right expertise, efficiency and scale.
FIGURE 11 Tiered price model
Price point in USD
6
5
4
3
2
1
0
1–100
101–200
201–500 501–1000
1000+
Thousands of subscribers
BOX C  
Legend for
Figure 12
FIGURE 12 Service evolution
Capture
aaS – as a service
BusCom –
business
communication
RCS – Rich
Communication
Suite
UC –unified
communication
VisualCom –
visual
communication
WebRTC – Refers
to standardization
for real-time
browser
capabilities.
VoLTE-aaS
RCS-aaS
Mobile
Grow
BusCom-aaS
VisualCom-aaS
UC-aaS
Enterprise
Innovate
WebRTC
Service
enablement
Cable/internet
References
1. Ericsson, February 2014, White Paper, The real-time cloud – combining cloud,
NFV and service provider SDN, available at:
http://www.ericsson.com/news/140220-the-real-time-cloud_244099438_c
2. European Commission, Digital agenda for Europe, Roaming, available at:
https://ec.europa.eu/digital-agenda/en/roaming
3. DMTF, Open Virtualization Format, available at:
http://www.dmtf.org/standards/ovf
Additional reading
ETSI, Network Functions Virtualisation, available at: http://www.etsi.org/
technologies-clusters/technologies/nfv
The author bios for Bart Jellema and
Marc Vorwerk can be found on page 19
E R I C S S O N R E V I E W • 2/2014
Connectivity for billions of things
12
Capillary networks – a smart
way to get things connected
A capillary network is a local network that uses short-range radio-access technologies to provide
groups of devices with connectivity. By leveraging the key capabilities of cellular networks – ubiquity,
integrated security, network management and advanced backhaul connectivity – capillary networks
will become a key enabler of the Networked Society.
JOAC H I M S AC H S , N IC K L A S BE I JA R , P E R E L M DA H L , JA N M E L E N, F R A NC E S CO M I L I TA NO A N D PAT R I K S A L M E L A
People and businesses
everywhere are becoming
increasingly dependent on the
digital platform. Computing and
communication are spreading
into every facet of life with ICT
functionality providing a way
to manage and operate assets,
infrastructure, and commercial
processes more efficiently. The
broad reach of ICT is at the heart
of the Networked Society, in
which everything will become
connected wherever connectivity
provides added value1,2 .
Ubiquitous connectivity and the
Networked Society
Connectivity in the Networked Society
is about increasing efficiency, doing
more with existing resources, providing services to more people, reducing
the need for additional physical infrastructure, and developing new services
that go beyond human interaction. For
example, smart agricultural systems
monitor livestock and crops so that irrigation, fertilization, feeding and water
levels can be automatically controlled,
which ensures that crops and livestock
remain healthy and resources are used
wisely. In smart health care, patients
and the elderly can get assistance
through remote monitoring – again
using resources in an intelligent way
– which improves the reach of health
care services, reduces the need for, say,
physical day clinics and cuts the need for
patients to travel.
As a whole, communication is progressively shifting from being humancentric to catering for things as well as
people. The world is moving toward
machine-type communication (MTC),
where anything from a smart device
to a cereal packet will be connected; a
shift that is to some extent illustrated
by the explosive growth of the Internet
of Things (IoT).
However, the requirements created
by object-to-object communication are
quite different from those of current
systems – which have primarily been
built for people and systems to communicate with each other. In scenarios where objects communicate with
each other, some use cases require battery-operated devices; therefore, low
energy consumption is vital. Barebones device architecture is essential
for mass deployment; typically the data
rate requirements for small devices are
low, and the cost of connectivity needs
to be minimal when billions of devices
are involved. Meeting all of these new
BOX A Terms and abbreviations
CoAP
EGPRS
eSIM
GBA
IoT
Constrained Application Protocol
enhanced general packet radio service
embedded SIM card
Generic Bootstrapping Architecture
Internet of Things
E R I C S S O N R E V I E W • 2/2014
MTC
machine-type communication
M2Mmachine-to-machine
OSPF
Open Shortest Path First
SLA
Service Level Agreement
TLS
transport layer security
requirements is a prerequisite for the
MTC business case.
Cellular communication technologies are being enhanced to meet these
new service requirements3,4. The powersave mode for example, introduced in
the most recent release (Rel‑12) of LTE,
allows a sensor that sends hourly reports
to run on two AA batteries for more
than 10 years, and simplified signaling
procedures can provide additional battery savings5. Rel-12 also introduces a
new LTE device category, which allows
LTE modems for connected devices to be
significantly less complex and cheaper
than they are today – the LTE features
proposed in 3GPP reach complexity levels below those of a 2G EGPRS modem6.
In addition, 3GPP has identified ways to
increase the coverage of LTE by 15-20dB.
This extension helps to reach devices in
remote or challenging locations, like a
smart meter in a basement 6.
Capillary networks and the shortrange communications technologies
that enable them are another key development in the Networked Society: they
play an important role providing connectivity for billions of devices in many
use cases. Examples of the technologies
include Bluetooth Low Energy, IEEE
802.15.4, and IEEE 802.11ah.
This article gives an overview of the
significant functionality that is needed
to connect capillary networks, including how to automatically configure and
manage them, and how to provide endto-end connectivity in a secure manner.
Capillary networks
The beauty of short-range radio technologies lies in their ability to provide connectivity efficiently to devices within a
13
specific local area. Typically, these local
– or capillary – networks need to be connected to the edge of a communication
infrastructure to, for example, reach
service functions that are hosted somewhere on the internet or in a cloud.
Connecting a capillary network to the
global communication infrastructure
can be achieved through a cellular network, which can be a wide-area network
or an indoor cellular solution. The gateway between the cellular network and
the capillary network acts just like any
other user equipment.
The architecture, illustrated in
Figure 1, comprises three domains: the
capillary connectivity domain, the cellular connectivity domain, and the data
domain. The first two domains span the
nodes that provide connectivity in the
capillary network and in the cellular
network respectively. The data domain
spans the nodes that provide data processing functionality for a desired service. These nodes are primarily the
connected devices themselves, as they
generate and use service data though
an intermediate node, which like a capillary gateway, would also be included in
the data domain if it provides data processing functionality (for example, if it
acts as a CoAP ­mirror server).
All three domains are independent
from a security perspective, and so
end-to-end security can be provided by
linking security relationships in the different domains to one another.
The ownership roles and business
scenarios for each domain may differ
from one case to the next. For example, to monitor the building sensors of a
real estate company, a cellular operator
might operate a wide-area network and
possibly an indoor cellular network, as
well as owning and managing the capillary network that provides the sensors
with connectivity. The same operator
may also own and manage the services
provided by the data domain and, if so,
would be in control of all three domains.
Alternatively, the real estate company
might own the capillary network, and
partner with an operator for connectivity and provision of the data domain. Or
the real estate company might own and
manage both the capillary network and
the data domain with the operator providing connectivity. In all of these scenarios, different service agreements are
FIGURE 1 System architecture for capillary network connectivity
Cellular access
Capillary network
Mobile
network
Connected
devices
M2M/IoT
cloud
Capillary gateway
Data domain
Capillary connectivity domain
Cellular connectivity domain
needed to cover the interfaces between
the domains, specifying what functionality will be provided.
Like most telecom networks, a capillary network needs a backhaul connection, which is best provided by a cellular
network. Their quasi-ubiquitous coverage allows backhaul connectivity to be
provided practically anywhere; simply
and, more significantly, without installation of additional network equipment.
Factoring in that a capillary network
might be on the move, as is the case for
monitoring goods in transit, leads to the
natural conclusion that cellular is an
excellent choice for backhaul.
In large-scale deployments, some
devices will connect through a capillary gateway, while others will connect to the cellular network directly.
Regardless of how connectivity is provided, the bootstrapping and management mechanisms used should be
homogeneous to reduce implementation complexity and improve usability.
Smart capillary gateway selection
Ideally, any service provider should
be able to deploy a capillary network,
including device and gateway configuration. For this to be possible, deployment
needs to be simple and use basic rules
– circumventing the need for in-depth
network planning. To achieve this, a
way to automatically configure connectivity is needed.
When deploying a capillary network,
a sufficient number of capillary gateways need to be installed to provide a
satisfactory level of local connectivity.
Doing so should result in a certain level
of connectivity redundancy – a device
can get connected through several different gateways. Some systems (such as
electricity meter monitoring) need to be
in operation for years at a time, during
which the surrounding environment
may change; nodes may fail, additional
network elements may be added, and
even the surrounding physical infrastructure can change. But, by allowing
the capillary network configuration to
change, some slack in maintaining constant connectivity is built into the system, which allows it to adapt over time.
The key to maintaining connectivity and building flexibility into connected systems lies in optimal gateway
selection. The decision-making process
– what gateway a device chooses for connectivity – needs to be fully automated
and take into consideration a number
of network and gateway properties.
Network parameters – such as the
E R I C S S O N R E V I E W • 2/2014
Connectivity for billions of things
14
quality of the cellular radio link and
the load in the cellular cell that a gateway is connected to – fluctuate, and so
a given capillary gateway will provide
different levels of backhaul connectivity at different times. Other considerations, like the amount of power a
battery-operated gateway has left, have
an impact on which gateway is optimal for a given device at a specific point
in time. Consequently, optimal gateway selection should not be designed
to balance load alone, but also to minimize delays, maximize availability
and conserve power. The gateway selection mechanism should support device
reallocation to another gateway when
the properties or the connectivity to a
gateway change. By designing gateway
selection to be smart, flexibility in connectivity is inbuilt, allowing systems
to continue to function as the environments around them evolve.
As illustrated in Figure 2, gateway selection relies on three different
types of information: connectivity, constraints and policy.
Connectivity information describes
the dynamic radio connectivity
between devices and gateways. Devices
typically detect connectivity by listening to the beacon signals that gateways
transmit. Some capillary short-range
radio technologies allow connectivity
to be detected by the gateway.
Constraint information describes the
dynamic and static properties of the network and the gateways that are included
in the selection process. Properties such
as battery level, load level (which can be
described by the number of connected
devices per gateway), support for QoS,
cost of use, and sleep schedule are all
included. The cellular backhaul connectivity of a gateway, such as link quality, can also be included, and future
enhancements might include properties such as cell load – obtained from the
management system of the cellular network. Devices may provide additional
constraint information, such as device
type, battery level, QoS requirements
and capillary network signal strength.
Policy information determines the
goal of gateway selection. A policy might
be a set of weightings or priorities that
determine how the various constraint
parameters affect the best choice of
gateway. Policy information may also
E R I C S S O N R E V I E W • 2/2014
include requirements set by the management system, such as allowing certain types of device to always connect
to given gateways. Policies are static and
are defined by network management.
The process of gateway selection
includes the following phases:
the information regarding connectivity,
constraints, and policy is gathered by
the element making the selection;
the gateway selection algorithm applies
the policies to the constraints while
taking connectivity into consideration
and determines the optimal gateway;
once a gateway has been selected for
each device, the selection is
implemented, which may imply that a
device needs to switch gateway; and
when a device moves to another
gateway, new routes to the device must
be set up in the cellular network so that
the incoming traffic is routed correctly.
The selection process can be controlled
at various locations in the network. The
location of control in turn affects the
need to transport information concerning constraints, policies and connectivity to the control point and to signal the
selection to devices.
If the control point is located in the
connected device, the device performs
the selection autonomously through
local computation based on information
sent by the gateway. As devices have just
a local view of the network, it may not
always be possible to optimize resources
globally and balance load across a group
of gateways.
If the control point is located in the
capillary gateways, the gateways need to
communicate with each other and run
the selection algorithm in a distributed
manner. This implies that gateways are
either connected via the capillary network, via the mobile network or via a
third network such as Wi-Fi, and use a
common protocol, like OSPF, for data
distribution. The main challenge here is
to reach convergence quickly and avoid
unnecessary iteration due to changes
in topology.
Alternatively the control point could
be a single node in the network that collects the entire set of available information. This centralized method enables
resource usage to be optimized globally
across the entire network. However, it
increases communication needs, as it
requires all of the capillary gateways to
communicate with a single point.
Managing QoS across domains
The QoS requirements for machinetype communication are typically different from those used for traditional
multimedia communication in terms
of bandwidth, latency and jitter. For
MTC, the requirement is often for guaranteed network connectivity with a
minimum throughput, and some use
cases may include stricter constraints
for extremely low latency.
For example, a sensor should be able
to reliably transmit an alarm within a
specified period of time after the detection of an anomaly – even if the network
is congested. To achieve this, low latencies are needed for real-time monitoring and control, while the bandwidth
requirements for this type of scenario
tend to be low. That said, QoS requirements for machine-type communication can vary tremendously from one
service to another. In some cases, like
surveillance, the QoS requirements are
comparable to those of personal multimedia communication.
QoS needs to be provided end-to-end.
So for the capillary network case, the
distinct QoS methods of both the shortrange network and the cellular network need to be considered. Each type of
short-range radio technology provides
different methods for QoS, which can
be divided into two main groups: prioritized packet transmission (for example,
in 802.11) and bandwidth reservation
(for example, in 802.15.4 and Bluetooth
Low Energy). As short-range technologies work in unlicensed spectrum, the
level of interference at any given time is
uncertain, which limits the level of QoS
that can be guaranteed. QoS methods
for the cellular networks that provide
connectivity, however, are well established and are based on traffic separation with customized traffic handling.
To provide QoS end-to-end, a bridge
is needed between the QoS domains of
the capillary and cellular networks. This
bridge specifies how traffic from one
domain (through a domain specific QoS
treatment) is mapped to a specific QoS
level in the other. The specifics of the
QoS bridge are determined in a Service
Level Agreement (SLA) established
between the providers of the capillary
15
FIGURE 2 Smart capillary gateway selection
3. Policies
Capillary
gateway
selection
Capillary
gateway
selection
1. Constraints
New communication path
4. (Re-) select gateway and
control communication path
2. Radio connectivity
Mobile
network
Mobile
network
M2M/IoT
cloud
Connected
devices
Capillary
gateways
network domain and the cellular connectivity domain, or between the service owner (in the data domain) and the
connectivity domain providers.
Security for connected devices
The devices deployed in capillary networks are likely to vary significantly in
terms of size, computational resources,
power consumption and energy source.
This variation makes implementing
and deploying security measures challenging. Security in capillary networks,
or within MTC in general, does not follow a one-size-fits-all model because
the constrained devices in the capillary
network are just that: constrained. It is
probably not possible to apply a generic
security solution: even if such a solution
ensures security in the most demanding
of scenarios, highly- constrained devices
will probably not have the resources to
implement it. What is needed is a security solution that fulfills the security
requirements of the use case at hand.
For example, a temperature sensor installed in a home is unlikely to
have the same strict security requirements as, say, a pacemaker or a sensor
in a power plant. A successful attack
on any one of these three use cases is
likely to yield drastically different consequences. So risk needs to be assessed
in the development of security requirements for the specific scenario, which
Old communication path
M2M/IoT
cloud
Connected
devices
Capillary
gateways
in turn determines what security solutions are suitable. The choice of a suitable security solution may then impact
the choice of device hardware, as it
needs to be capable of implementing
the selected security solution.
For end-to-end protection of traffic between authenticated end-points,
widely used security mechanisms such
as TLS would improve interoperability between constrained devices and
services that are already deployed. In
some cases, there might be a need for
more optimized security solutions to
be deployed, such as by using a protocol
that entails fewer round-trips or incurs
less overhead than legacy solutions.
Identification
When a device is installed in a capillary network, in most cases it needs to
possess some credentials – that is to say
an identity and something it can use
to prove it owns the identity, such as a
key. Typical solutions include public key
certificates, raw public keys or a shared
secret. With its stored credentials, the
device needs to be able to authenticate
itself to the services it wants to use –
such as a management portal through
which the device is managed, a data
aggregation service where the device
stores its data, as well as the capillary
gateway, which provides the device with
global connectivity.
One way to implement device identification and credentials is to use the
same method used in 3GPP networks
– basically the 3GPP subscription credentials. The subscription identity and
a shared secret that can be used for
authentication in 3GPP networks are
stored on the SIM card of the device. In
addition to using the credentials to get
network access, they can also be used
for authenticating the device to various services in the network. This can
be done using the 3GPP-standardized
Generic Bootstrapping Architecture
(GBA). For MTC scenarios, GBA is a good
solution, as it provides strong identification and communication security
without requiring any user interaction
or configuration at the device end; the
security is based on the 3GPP credentials stored in a tamper-resistant environment, to which not even the user has
direct access.
To apply GBA, first of all the device
needs to have 3GPP credentials; and
then the 3GPP network, the desired service as well as the device itself all need
to support GBA. Unfortunately, many
capillary network devices do not possess 3GPP credentials, which limits the
use of GBA to capillary gateways. In such
cases, the gateway can provide GBAbased authentication and security for
services on behalf of the entire capillary
network, but device authentication
E R I C S S O N R E V I E W • 2/2014
Connectivity for billions of things
16
FIGURE 3 Security domains – bootstrapping and management
Capillary gateway security domain
Capillary device security domain
Connectivity security domain
Data security domain
Capillary network
Mobile
network
Connected
devices
M2M/IoT
cloud
Capillary gateway
Trust/business relationship
GBA-based security
Alternative
End-to-end security solution
still needs to be performed between
the device and the service.
Security domains
Capillary networks have two distinct
security domains, as illustrated in
Figure 3: the capillary devices and the
capillary gateway that provides widearea connectivity. The security domain
for devices can further be split into connectivity and data domains. The data
domain incorporates the device and
the services it uses, such as management and data storage, and the connectivity domain handles the interaction
between the device and the capillary
gateway.
The security domain for the capillary
gateway is based on the 3GPP subscription and the security that the subscription credentials can provide for access
services and 3GPP-aware services; for
example, through the use of GBA.
The two security domains intersect
at the capillary gateway; there is a need
for mutual trust and communication
security between the device and the
E R I C S S O N R E V I E W • 2/2014
gateway. At this intersection there is an
opportunity to apply the strong identification and security features of the 3GPP
network for the benefit of the capillary
device. If strong trust exists between
the device and the capillary gateway,
the security domains can be partially
merged to provide the device with 3GPPbased security for the GBA-enabled services it uses.
Bootstrapping
When a device is switched on or wakes
up, it may be able to connect to a number of capillary gateways, possibly provided by different gateway operators.
The device needs to know which gateway it has a valid association with and
which it can trust. Once global connectivity has been established, the device
also needs to know which services to
connect to. Capillary devices will be
deployed in the thousands, and as a consequence of their bare-boned architecture, they do not tend to be designed
with easy-to-use user interfaces. Manual
configuration of massive numbers of
capillary devices has the potential to
be extremely time consuming, which
could cause costs to rise.
Bootstrapping devices to their services using a bootstrap server is one
way of automating configuration and
avoiding the manual overhead. Such a
service, which could be operated by the
device manufacturer, would ensure that
the device is redirected to the selected
management service of the device
owner. During the manufacturing
process, devices can be pre-configured
with information about the bootstrap
server, such as how to reach it and how
to authenticate it. When switched on or
upon waking up, the device will connect
to the bootstrap server, which helps it to
find its current home.
If a device gets corrupted, or for
some reason resets itself, it can – once
rebooted – use the bootstrap server to
reach its current management portal.
From the management portal, either
the device owner or an assigned manager can configure the device with the
services it should use – and possibly even
provide the service specific credentials
to the device. This approach removes
the need to individually configure each
device, and can instead provide a centralized point for managing all devices,
possibly via batch management.
The ability to remotely manage
devices becomes significant when, for
example, 3GPP subscription information needs to be updated in thousands
of deployed devices. Today, 3GPP credentials tend to be stored on a SIM card,
and updating this information typically requires replacing the SIM card
itself. Embedded SIM cards (eSIM) and
SIM-less alternatives are now being
researched. While eSIM is a more MTCfriendly option, as it allows for remote
management of subscription information, SIM-less is of most benefit to constrained devices, to which adding a SIM
is an issue simply because they tend to
be quite small.
Network management
A range of tasks, such as ensuring automatic configuration and connectivity
– for devices connected through a capillary network – are fulfilled by network
management. In addition, network
management needs to establish access
control restrictions and data treatment
17
rules for QoS based on SLAs, subscriptions and security policies. In addition,
a service provider should be able to use
the management function to adapt service policies and add or remove devices.
By nature, connected devices are rudimentary when it comes to manual interaction capabilities. Additionally, the fact
that service providers tend to have no
field personnel for device management
implies that a remote management and
configuration interface is needed to be
able to interact with deployed devices.
Network management of connected
devices in capillary networks poses new
challenges compared with, for example,
the management of cellular networks.
This is partly due to the vast number of
devices, which are orders of magnitude
larger than the number of elements
handled by today’s network management systems. Instead of handling
devices as individual nodes, economy of
scale can be achieved by handling them
in groups that use policies and managed
parameters that are more abstract and
also fewer in number.
Consider the case of a service provider
that wants to reduce costs by replacing sensor batteries less frequently.
To achieve this, the service provider
increases the life length policy of the
node in the management system. The
management system interprets this policy and sets the reporting frequency to
every two hours, instead of every hour,
for a group of sensors in a particular geographical region.
Connected devices will often be battery powered, and so all operations,
including management, need to be
energy optimized to reduce the impact
on battery usage. Additionally, connected devices tend to sleep during
extended periods of time, and so management operations cannot be expected
to provide results instantly, but only
after the device wakes up.
A significant challenge for network
management is the provision of full endto-end scope, an issue that is particularly evident when different domains
in the end-to-end chain are provided
by different business entities – as discussed and indicated in Figure 1. Based
on analysis of the connectivity information provided just by the devices,
the connectivity state can only be estimated at a high level, extracted from
the information available at each end
of the communication path. Estimating
the connectivity in this way can lead
to a significant overhead to obtain and
maintain such information; it is also
limits the configuration possibilities of
the connectivity layer.
The best way to overcome this limitation is to interconnect the network
management systems in the different domains. In this way, connectivity information from the nodes along
the communication path, between the
end points, can also be included. If the
domains are operated by separate entities, this can be achieved through SLAs
specifying the usage and exchange
of information. The resulting crossdomain management provides end-toend management opportunities. For
example, QoS in both the capillary and
the 3GPP domains can be matched, and
alarms from both domains can be correlated to pinpoint faults.
Summary
As the Networked Society starts to take
shape, a vast range of devices, objects
and systems will be connected, creating
the Internet of Things (IoT). Within this
context, cellular networks have a significant role to play as connectivity providers, to which some things will connect
directly, and another significant portion will connect using short-range
radio technologies through a capillary
network.
Cellular networks can provide global
connectivity both outdoors and indoors
by connecting capillary networks
through special gateways. However,
achieving this will require some new
functionality.
Due to the massive numbers of connected things, functionalities – such as
self-configuring connectivity management and automated gateway selection
– are critical for providing everything
in the capillary network with a reliable
connection.
To ensure that communication
remains secure and trustworthy, a security bridge is needed between the capillary and the cellular domains. With this
functionality in place, a future network
can provide optimized connectivity for
all connected things anywhere no matter how they are connected.
References
1. Morgan Stanley, April 2014, Blue Paper, The ‘Internet of Things’ Is Now:
Connecting The Real Economy, available at:
http://www.morganstanley.com/views/perspectives/
2. J. Höller, V. Tsiatsis, C. Mulligan, S Avesand, S. Karnouskos, D. Boyle, 1st edition,
2014, From Machine-to-Machine to the Internet of Things: Introduction to a New
Age of Intelligence, Elsevier, available at:
http://www.ericsson.com/article/from_m2m_to_iot_2026626967_c
3. Alcatel Lucent, Ericsson, Huawei, Neul, NSN, Sony, TU Dresden, u-blox, Verizon
Wireless, White Paper, March 2014, A Choice of Future m2m Access Technologies
for Mobile Network Operators, available at: http://www.cambridgewireless.
co.uk/docs/Cellular%20IoT%20White%20Paper.pdf
4. Ericsson, NSN, April 2014, LTE Evolution for Cellular IoT, available at: http://www.
cambridgewireless.co.uk/docs/LTE%20Evolution%20for%20Cellular%20
IoT%2010.04.14.pdf
5. Emerging Telecommunications Technologies, April 2014, T. Tirronen, A. Larmo, J.
Sachs, B. Lindoff, N. Wiberg, Machine-to-machine communication with long-term
evolution with reduced device energy consumption, available at:
http://onlinelibrary.wiley.com/doi/10.1002/ett.2643/abstract
6. 3GPP, TR 36.888, June 2013, Study on provision of low-cost Machine-Type
Communications (MTC) User Equipments (UEs) based on LTE, available at:
http://www.3gpp.org/DynaReport/36888.htm
E R I C S S O N R E V I E W • 2/2014
Connectivity for billions of things
18
Joachim Sachs
is a principal researcher at Ericsson Research. He joined Ericsson in
1997 and has worked on a variety of topics in the area of wireless
communication systems. He holds a diploma in electrical engineering
from Aachen University (RWTH), and a doctorate in electrical
engineering from the Technical University of Berlin, Germany. Since 1995 he has been
active in the IEEE and the German VDE Information Technology Society (ITG), where
he is currently co-chair of the technical committee on communication.
Nicklas Beijar
is a guest researcher at Ericsson Research in the Cloud Technologies
research area. He joined Ericsson in 2013 to work with the Internet of
Things and, in particular, he has been working on the capillary network
prototype demonstrated at Mobile Word Congress 2014. His current
focus is on cloud-based solutions supporting the IoT. He holds a D.Sc. in networking
technology from Aalto University and an M.Sc. from the Helsinki University of
Technology, both in Finland.
Per Elmdahl
is a senior researcher at Wireless Access Networks, Ericsson Research.
He holds an M.Sc. in computer science and technology from Linköping
University, Sweden. He joined Ericsson in 1990 researching network
management and network security. He served as an Ericsson 3GPP SA5
delegate for seven years, working on network management. While his interest in the
IoT began privately, he has worked on the subject professionally for the last two years,
specifically on network management and Bluetooth Low Energy.
Jan Melen
is a master researcher at Ericsson Research in the Services Media and
Network Features research area. He joined Ericsson in 1997 and has
worked with several 3GPP and IP related technologies. He studied at the
electrical engineering department at Helsinki University of Technology,
Finland. He has been involved in several EU projects, IETF and 3GPP standardization.
He has been leading the IoT related research project at Ericsson Research since 2011.
Francesco Militano
is an experienced researcher at Ericsson Research in the Wireless
Access Networks department. He joined Ericsson in 2011 to work with
radio architecture and protocols. At present, he is investigating the field
of M2M communications with LTE and capillary networks. He holds an
M.Sc. in telecommunications engineering from University of Siena, Italy, and a postgraduate degree in networks innovation and ICT sector services from the Polytechnic
University of Turin (Politecnico di Torino), Italy.
Patrik Salmela
is a senior researcher at Ericsson Research focusing on security. He
joined Ericsson in 2003 to work for Ericsson Network Security and
moved one year later to Ericsson Research, where he focused for several
years on the Host Identity Protocol. He has since been working on
security topics related to 3GPP, Deep Packet Inspection, and most recently, the
Internet of Things. He holds an M.Sc. in communications engineering from Helsinki
University of Technology, Finland.
E R I C S S O N R E V I E W • 2/2014
19
Authors
Communications as a
cloud service: a new take
on telecoms
Pages 4-11
Marc Vorwerk
joined Ericsson in 2000.
Today, he is a senior
specialist for cloud
computing, and has
previously worked on multi-access, IMS
and media-plane management
research – developing early prototypes
and participating in European research
projects. He began utilizing
virtualization and cloud over six years
ago, and has been an evangelist within
Ericsson to promote the benefits of
these technologies. Today as a senior
specialist he is a team leader, an
innovation event presenter and
provides customer-engagement
support. He holds an M.Sc. in electrical
engineering from RWTH Aachen
University, Germany.
Bart Jellema
joined Ericsson in 1989.
He has held several
system and product
management roles in
Canada, Germany and the Netherlands.
He currently works with the core
networks architecture and technology
team in the area of cloud and NFV, and
is involved in the establishment of
Ericsson’s new global ICT centers. He
has been active in standardization,
holds several patents and is a speaker
for Ericsson at innovation events. He
holds a B.Sc. in electrical engineering
from the University of Applied
Sciences, Eindhoven, the Netherlands.
E R I C S S O N R E V I E W • 2/2014
Can it be trusted?
20
Trusted computing for
infrastructure
The Networked Society is built on a complex and intricate infrastructure that brings distributed
services, data processing and communication together, combining them into an innovative and
more meaningful set of services for people, business and society. But combining services in such an
advanced way creates new requirements in terms of trust. Trusted computing technologies will play
a crucial role in meeting the security expectations of users, regulators and infrastructure owners.
M I K A E L E R I K S S ON, M A K A N P OU R Z A N DI A N D BE N S M E E T S
Today’s industries are in
transformation and ICT is
changing the game. New
applications built from a
combination of services,
communication and virtualization
are being rolled out daily,
indicating that the Networked
Society is becoming reality.
Communication is transitioning from
a person-to-person model to a system
where people, objects and things use
fixed and mobile connections to communicate on an anything-to-anything,
anywhere and anytime basis. But even
though people and businesses are beginning to use and benefit from a wide
range of innovative applications, the
potentially massive benefits that can
be gained by combining modern computing, web services and mobile communication have yet to be realized.
As we progress deeper into the
Networked Society, people, systems
and businesses will become ever more
dependent on an increasingly wider
range of internet and connected services. And so the fabric of the Networked
Society needs to be built on solutions
that are inherently secure, socially
acceptable and reliable from a technical point of view.
Modern internet services rely on
web and cloud technology, and as such
they are no longer independent packages with in-built security, but are constructed through the combination
and reuse of other services distributed
across the web. This creates new issues
BOX A Terms and abbreviations
BIOS basic input/output system
CBA Component Based Architecture
DoS denial-of-service
DRM Digital Rights Management
DRTM dynamic RTM
HE Homomorphic Encryption
MME Mobility Management Entity
OS operating system
PKI public key infrastructure
ROM read-only memory
RoT Root of Trust
RTM RoT for measurement
RTR RoT for reporting
RTS RoT for storage
SDN
software-defined networking
SGSN Serving GPRS Support Node
SGSN-MME Network node combining SGSN and MME functions
E R I C S S O N R E V I E W • 2/2014
SGX Software Guard Extensions
SICS Swedish Institute of Computer Science
SLA Service Level Agreement
SRTM static RTM
SSLA Security Service Level Agreement
TCB trusted computing base
TCG Trusted Computing Group
TEE Trusted Execution Environment
TLS Transport Layer Security
TPM Trusted Platform Module
TXT Trusted eXecution Technology
UEFI
Unified Extensible Firmware Interface
VM virtual machine
VMM virtual machine manager (hypervisor)
vTPM virtual TPM
in terms of security. One of the most
fundamental of these issues is securing processing in the communication
infrastructure so that it can be trusted.
Solving this issue is a prerequisite for
building trust relationships into a network fabric for data communication
and cloud computation. The red arrows
in Figure 1 illustrate possible trust relationships in such a network fabric that
connects servers, data centers, controllers, sensors, management services, and
user devices.
Trusted computing concepts
Users and owners of processing nodes
use trusted computing to assess the certainty of one or several of the following
aspects:
what the processing nodes do;
how nodes protect themselves against
threats; and
who is controlling the nodes.
This includes determining where data is
stored and processed – which can be significant when legal or business requirements related to data handling need to
be met.
This article presents an overview of
the technical solutions and approaches
for implementing trusted computing in
a telecommunications infrastructure.
Some of the solutions follow the concepts outlined in the Trusted Computing
Group (TCG) specifications. Together the
solutions described here enable what is
often referred to as a Trusted Execution
Environment (TEE), and with the addition of platform identities they provide
a means for secure access control and
management of platforms.
21
In this article, the term platform is
used to refer to the technical system for
computational processing, communication and storage entities; which can be
physical or virtual. The term infrastructure is used to refer to a wider concept,
normally consisting of a collection of
platforms and networks that is designed
to fulfill a certain purpose.
Ensuring that the implementation of
a technical system can be trusted calls
for assurance methodologies. How to
apply a security assurance methodology
to every stage of product development,
so that the implementation of a securityassurance product is in accordance with
agreed guidelines has been discussed in
a previous Ericsson Review article1.
A model for trust
The infrastructure, which is illustrated
in Figure 1, consists of servers, routers,
devices and their computational, communication and storage aspects. This
complex set of relationships can be redesigned using a cloud-based model – as
shown in Figure 2. While the cloud
model also consists of devices, access
nodes, routing units, storage, servers
and their respective management processes, the principles of trusted computing have been applied, and so the
building blocks of each entity include
trusted computing sub-functions.
Management functions govern the
behavior of the platforms through
a number of Security Service Level
Agreements (SSLAs). For example, an
SSLA might impose policies for booting, data protection or data processing.
Through a trustworthy component
known as Root of Trust (RoT), each entity
locally enforces and checks for SSLA
compliance. An RoT may be referred to
as a trusted computing base (TCB) or
trust anchor. It can be implemented
as a hardware component, or exposed
through a trusted virtual entity.
The RoT is one of the fundamental
concepts of trusted computing for providing protection in the cloud model
illustrated in Figure 2. Together with a
set of functions, an RoT is trusted by the
controlling software to behave in a predetermined way. The level of trust may
extend to external entities, like management functions, which interact
remotely with the RoT and contribute
to establishing a trustworthy system.
FIGURE 1 Examples of trust relationships in the Networked Society
Data center
Data center
management
Network
Network
management
Gateway
Routing
HSS
Server
management
Access
Device
management
How the terms trust and trustworthiness are interpreted can be quite complex. They may depend on the results
of an evaluation (such as Common
Criteria methodology for Information
Technology Security Evaluation1), or of a
proof, and may even depend on the reputation of the organization or enterprise
delivering the RoT. An RoT can provide
several functions, such as:
verification of data authenticity and
integrity;
provision and protection of secure
storage for secret keys;
secure reporting of specific machine
states; and
secure activation.
In turn, these functions allow features
such as boot integrity, transparent drive
encryption, identities, DRM protection,
and secure launch and migration of virtual machines (VMs) to be built.
The implementation of an RoT must
be able to guarantee a certain level of
assurance against modification. A good
example of this is the ROM firmware
that loads and verifies a program during a boot process. The TCG approach to
trusted computing relies on the interaction of three RoTs to guarantee protection from modification – each one with
a specific task (see Box C):
storage – the RoT for storage (RTS);
measurement – the RoT for
measurement (RTM); and
reporting – the RoT for reporting (RTR).
How these RoTs are implemented
is highly dependent on the Trusted
Platform Module (TPM) and the cryptographic keys that are used to secure
device hardware.
E R I C S S O N R E V I E W • 2/2014
Can it be trusted?
22
FIGURE 2 A trusted computing cloud model
Compute
(process)
Communicate
Storage
Run-time integrity, protection and privacy
Data integrity: at rest and in motion
Trusted compute initialization: boot integrity
Identity: personalization, provisioning
Measurement
The RoT for measurement – RTM – is
defined in the platform specification
and provides the means to measure the
platform state. It comes in two flavors:
static and dynamic – SRTM and DRTM,
respectively. Intel’s TXT, for example, is a
DRTM; it supports platform authenticity
attestation and assures that a platform
starts in a trusted environment. The
RTM is a crucial component for ensuring that a platform is in a trusted state.
In contrast to the reporting and storage
RoTs, the RTM resides outside the TPM –
see Box C. A DRTM can be used to bring
a platform into a trusted state while it
is up and running. Whereas the static
flavor starts out from a trusted point,
based on a fixed or immutable piece of
trusted code as part of the platform boot
process.
Chipset vendors and platform manufacturers decide what flavor the RTM
should be implemented in – static or
dynamic. The implementation of Intel’s
TXT, for example, includes many adaptations in the chipset, and even uses
Intel propriety code.
E R I C S S O N R E V I E W • 2/2014
A TPM is often implemented as a separate hardware component that acts as
a slave device. However, it can be virtualized, and in this case is often referred
to as a vTPM (see2, for example). To
implement an RoT, there are other solutions than strictly following the TCG
approach, such as those built using the
ARM TrustZone concept. TrustZone can
itself be used to implement an RoT as an
embedded TPM with the functions mentioned in Box C.
Business aspects
In the Networked Society, cloud computing and cloud-based storage will
be widely deployed. These technologies rely on a trustworthy network fabric; however, in a recent survey of the
Open Data Center Alliance, 66 percent
of the members stated that they are concerned about data security3. The upshot
of this has been a delay in the adoption
of cloud computing. Consequently, the
use of trusted computing in existing
and emerging cloud solutions is highly
desirable, as it will help to dispel the
fears associated with data security, lead
Management
SSLA
SSLA
SSLA
PKI
to increased service use and new business models, and create opportunities
for technological leadership.
Other business aspects influencing
trusted computing solutions include
requirements for scalability and elasticity of cloud computing, and the extent to
which processing will be self-governed.
In the cloud
Trusted computing in a cloud environment is a special case. Web services and
programmable routing technology
(SDN based) using infrastructures like
the one illustrated in Figure 1, will be
deployed on platforms that exploit virtualization. To ensure overall security
in the cloud, both the launch and the
operation of virtualized resources need
to be secure.
With respect to Figure 2, three
core features are essential for building trusted computing in a cloud
environment:
boot integrity – so that the hardware
platform can guarantee a trustworthy
RoT for the overall cloud environment;
secure management of VMs – to secure
the launch and migration of VMs in the
cloud environment; and
secure assessment of VMs – to attest
the security and trustworthiness of VMs
throughout their life cycles.
Boot integrity
To boot a platform in a trustworthy way,
a bootstrap process that originates from
an immutable entity – an RoT – must be
used. If the RoT provides proof of the
progress of the bootstrapping process
to the user in some transparent way, it
acts as a measurement RoT.
There are two main approaches to the
bootstrapping process: a verified boot or
a measured boot.
A verified boot actively attests each
component before it is loaded and executed. Using this approach, a platform
will either boot or fail, depending on
the outcome of the verification of each
component’s cryptographic signature.
Measured boot, on the other hand, is
passive. Here, each component is measured and progress reports are saved
into safe storage. The controlling process can then parse the recorded measurements securely and determine
whether to trust the platform or not.
Of the two approaches, only measured
23
boot complies with TCG; measurements
combined with attestation are referred
to as a trusted boot.
Both approaches can be used independently, or combined in a hybrid version
to extend the integrity of the boot to client applications – which is illustrated in
Figure 3. At Ericsson, ongoing work in
Component Based Architecture (CBA)
aims to establish a common approach
to boot solutions and signed software;
coordinating use in products.
Secure launch
Security-sensitive users need assurance
that their applications are running on a
trustworthy platform. Such a platform
provides a TEE and techniques for users
to attest and verify information about
the execution platform.
In some cases, clients may want to
receive an attestation directly from
the platform. To do this, users need to
be provided with a guaranteed level of
trust in hardware or the virtualization
layer during the initial VM launch, as
well as throughout the entire VM life
cycle – migration, cloning, suspension
and resumption.
To launch a VM in a secure way, the
security and trustworthiness of the
hardware platform and virtual layer
first need to be attested. For certain sensitive applications, like financial transactions or handling legal intercept, the
VM or the owner of the VM need to be
advised on the trustworthiness of the
hardware platform each time the hardware platform is changed – for example
following the migration, suspension or
resumption of a VM.
In a cloud environment, some additional security constraints may apply
to a VM launch. For example, due to the
risk of a side channel or a DoS attack,
some customers may require their virtual resources to be separated (not colocated) from any other customer’s
resources.
There are basically two of ways of
attesting a secure VM launch to clients:
the cloud provider can deploy the
trusted cloud and prove its
trustworthiness to the client; or
trustworthiness measurements can be
conveyed to the client – either by the
cloud provider or by an independent
trusted third party.
In the first approach, customers must
trust the cloud provider. The difficulty
with the second approach is the ability of a customer or trusted third party
to collect the trustworthiness evidence
related to the cloud providers – given
the dynamic nature of the cloud and
the diverse set of hardware, operating
systems (OSs) and VM managers (VMMs)
used. This task becomes even more complex because trustworthiness needs to
be reestablished and checked every
time a change occurs in the underlying
layers: hardware, OS, and VMM. It seems
inevitable that for the second approach
to work, cloud providers would have
to expose some, or all of their internal
hardware and software configuration,
including, say, hardware platform specifics, OS, VMM, and even configuration
information and IP addresses. This may
conflict with a cloud provider’s policy to
keep its internal architecture private.
The solution presented in Huebner
on Intel TXT4 is of the first type – based
on trust. Here, attestation is achieved
inside the cloud environment, and the
results are then provided to users.
The BIOS, OS, and hypervisor of the
hardware platform are measured, and
the results are sent to an attestation
server. The server in turn verifies their
trustworthiness by comparing them
against a known database of measurements. Following successful verification, the secure VM launch can then be
carried out.
When attestation is achieved through
the trust model, users cannot remotely
attest the hardware platform and consequently have to trust the cloud provider
and its attestation through SLAs.
To attest a secure VM launch using
the second approach – based on measurement – Ericsson security researchers have created a framework5,6 in
OpenStack to verify the trustworthiness of VM host system software
through remote attestation with a
trusted third party.
FIGURE 3 Hybrid boot process using an RoT for measurement
Platform
management
Proof – through signature
(for example)
Anti-malware software is
started before any third party
software
Login
Attestation
service
Client
Measurements
Third party
software/drivers
RTM
Kernel and
drivers
Anti-malware
software/drivers
Boot manager
firmware
Anti-malware
policy
UEFI boot
Boot policy
Client can fetch TPM
measurements of
client state
TPM
Measurements of components and
anti-malware software are recorded
in the TPM
E R I C S S O N R E V I E W • 2/2014
Can it be trusted?
24
FIGURE 4 Trusted computing attestation process
Open stack
BOX B Cloud
management
Trusted computing pool
4
1
2
3
Server
management
Secure migration
In a cloud environment, VM migration
is often necessary to optimize the use
of resources and ensure optimal power
consumption. This is a highly dynamic
process that depends on many factors,
including application needs, host loads,
traffic congestion, and hardware and
software failures. A secure VM migration ensures the security of the VM both
at rest and during the migration – guaranteeing the same level of trust before
and after. Similarly, cloud federation use
cases require interoperability guarantees among the different cloud service
providers. To achieve this, mechanisms
need to be in place to ensure the same
level of trust when a VM is migrated
from one cloud provider to another.
Migrating a VM can sometimes result
in a change of underlying hardware that
the VM is not aware of. This is significant, as the RoT function can depend
on both hardware and VMM (when it
comes to virtual TPM deployment for
VMs). Migrations are often performed
programmatically by cloud orchestration or management in a manner that is
transparent to the VM. So, cloud orchestration and management need to be
involved to choose the right physical
hosts and VMMs with adequate levels
of trust expressed in SSLAs to run VMs.
For regulation or auditing purposes,
preserving proof of trustworthiness of
E R I C S S O N R E V I E W • 2/2014
Scheduler
the platform needs to be provided for
security-sensitive applications. This use
case can be extended to a remote attestation of HW-VMM-VM to the tenant’s
auditor. There are two aspects related
to preserving trustworthiness:
ensure that the hardware and VMM after
the migration can be trusted to preserve
the same level of trust (trusted
computing base) for VM before and after
the migration; and
provide the same RoT functionality to a
VM before and after migration: for
example, protection and storage of
secret keys in a virtual TPM.
So far, secure VM migration has
received less attention than the secure
launch from both academia and industry. Despite this lack of interest, secure
VM migration is an essential part of the
overall secure life cycle of VMs, if satisfactory levels of security for applications
in the cloud are to be achieved.
Secure assessment
From a management point of view, the
platform needs to provide trustworthiness information and provide assurance
that it responds correctly to management commands. Remote assessment
of the platform state is of particular
importance to ensure that the launch
or migration of a virtual machine is carried out securely.
Trusted
computing
attestation
process
1) the Open
Attestation
Server
determines
a trusted
computing
pool;
2) cloud
management
requests new
workloads from
the scheduler;
3) the
scheduler
requests
the list of
trusted
computing
nodes in
the trusted
computing
pool; and
4) the workload
is initiated on
a computing
node inside
the trusted
computing
pool.
Obtaining assurance for every single functional aspect of the platform
and the services it hosts can be difficult.
Obtaining assurance for just a limited
set of functions can reduce the complexity of this task and be an acceptable
trade-off. Ideally, those aspects that have
security relevance should be expressed
in an agreement between the provider
and the user – typically detailed in an
SSLA, which might demand the support of remote assessment procedures.
For this, a platform should have a set of
mechanisms, like RTM coupled to RTR,
that allow a remote entity to securely
assess certain properties recorded by
monitoring capabilities of the platform’s local trustworthy subsystem. Yet
proper assurance methodologies have
to be applied to ensure that these mechanisms deliver what is needed without
any blind spots, which would result in a
false sense of security.
Implementation aspects
Standards
Although extensive academic work has
been carried out in the field of trusted
computing, only a few implementation standards exist for interoperable
trusted computing solutions. The TCG
has specified a framework and components for implementing trusted computing, which are used by chipset vendors
such as Intel and AMD. However, the
TCG specifications can result in varying
implementations by the different vendors, which is good, as different vendors
can optimize their solutions for different capabilities such as for performance
or for storage. While this flexibility is
advantageous, it also creates interoperability issues7. Flexibility has been
further increased in TPM 2.0 through
implementation and choice of cryptographic primitives. Currently, the TCG
specifications remain the most comprehensive standards for implementing RoTs.
Another important set of specifications has been issued by the
GlobalPlatform organization. Its TEE
specifications include architecture for
secure computation and a set of APIs.
Although these specifications provide
trusted computing for mobile devices,
they can also be used for infrastructure nodes such as base stations. How
the secured environment is actually
25
implemented is left to the discretion of
the hardware vendors and can be system-on-chip or a dedicated separate
component; ARM TrustZone is an example implementation of this technology.
As of mid-2014, the GlobalPlatform specifications do not address how a system
reaches a trustworthy state and how
trust properties can be asserted. With
this in mind, the GlobalPlatform and
TCG specifications complement each
other.
Hardware aspects
As illustrated by the Intel TXT implementation of the TCG DRTM concept,
several components in general purpose
chipsets must be modified to achieve
the needed protection. Similarly, the
protection provided by TrustZone
affects the ARM core as well as its subsystems. This level of invasiveness
results in hardware vendors sticking
to their chosen approach to trusted
computing, and changes to functionality tend to be implemented in a stepwise fashion. Intel and AMD have been
using TCG functionality, and ARM has
pursued its TrustZone concept and
announced cooperation with AMD.
Unfortunately, the TCG specifications
do not really cover the aspects of isolation of execution. To fill this gap, Intel
introduced the SGX concept, which is a
set of new CPU instructions that applications can use to set aside private regions
of code and data.
Isolation during execution is an
important principle, and future hardware will have more functionality to
improve isolation and control of the execution environments. The SGX concept
also supports attestation and integrity
protection, as well as cryptographic
binding operations per private region.
Homomorphic Encryption
In some (cloud) processing cases, it might
be possible to apply what is referred to
as Homomorphic Encryption (HE) as
an alternative to applying stringent
secrecy demands on processing nodes.
Current research in this subject and
similar techniques appear to be promising – leading to reasonably fast cloudbased processing of secret (encrypted)
data for certain operations without
needing to make the data available in
clear text to the processing node.
However, HE is a rather undeveloped technology; it only solves certain aspects of trusted computing, and
involves a level of computational complexity that is, generally speaking, still
too high. It may, however, become a
complementary technique for trusted
computing. If that happens, hardware
support for HE operations will likely
find its way onto server chipsets.
Examples of platform security
In cooperation with the Swedish
Institute of Computer Science (SICS),
Ericsson Research has modified
OpenStack to use a TPM for secure VM
launch and migration. A trusted third
party was used for collecting and sending trustworthy information and control. Part of the solution has been used
in a cloud-based test-bed setup for a
regional health care provider in southern Sweden.
Ericsson security researchers have
also implemented solutions for cloudbased protection of persistent storage8.
Generally speaking, secure VM launch
and migration are finding their way into
OpenStack.
The coming release of the Ericsson
SGSN-MME node is another example of
how trusted computing has been implemented using TPM technology. Beyond
the functionality discussed above, the
TPM is used for secure storage of PKI
credentials. These credentials are used
for TLS connections and for encryption
of sensitive data. Like other telco nodes,
the SGSN-MME has high-availability
requirements, which calls for the use
of hardware redundancy and efficient
maintenance procedures. As the TCG
specifications do not address such use
cases, special care must be taken when
deploying TPMs in such a setting: production, personalization, rollout, and
maintenance support have to be implemented before any of the trusted computing features can be enabled.
Conclusion
Ericsson recognizes that trusted computing is a technical approach that will
enable secure infrastructures and services for the Networked Society. As the
use of virtualization technologies and
the cloud increases, maintaining trust is
essential. In connection with the cloud,
the use of a virtual trusted platform
model as an RoT for different virtual
machines has received some attention from both academia and industry. Despite this, further development
is required to address issues related to
establishment of trust models, trusted
evidence collection, and real-time and
dynamic attestation. Ericsson Research
is active in this field and cooperates with
Ericsson business units to incorporate
such security solutions into products.
BOX C Three main TPM tasks
TPM
The TPM is responsible for protecting secret keys and sensitive functions. The bulk
of the TPM’s data is stored outside the TPM in so-called blobs. The RTS provides
confidentiality and integrity protection for these blobs. The RTR is responsible for:
reporting platform
configurations;
protecting reported
values;
providing a function for
attesting to reported
values; and
establishing platform
identities.
The interaction between the RTR and RTS relates to the responsibility for protecting
measurement digests. The term measurement has a specific meaning in TCG and can
be understood as verification in relation to RTM functions.
RTR
RTS
Protected
capabilities
Shielded
locations
RTM
E R I C S S O N R E V I E W • 2/2014
Can it be trusted?
26
Mikael Eriksson
is a security architect at Business Unit Cloud & IP. He holds an M.Sc. in
data and image communication from the Institute of Technology:
Linköping University, Sweden. He joined Ericsson in 2009 to work with
mobile broadband platforms after an 18-year career as a consultant,
mostly in embedded systems. Since 2012, he has been with the Packet Core Unit,
working on adaptation of security technology in mobile networks infrastructure. He is
currently the study leader of a boot integrity integration project of Ericsson platforms.
Makan Pourzandi
works at Ericsson Security Research in Montreal, Canada. He has more
than 15 years of experience in security for telecom systems, cloud and
distributed security and software security. He holds a Ph.D. in parallel
computing and distributed systems from the Université Claude Bernard,
Lyon, France, and an M.Sc. in parallel processing from École Normale Supérieure
(ENS) de Lyon, France.
Ben Smeets
is an expert in security systems and data compression at Ericsson
Research in Lund, Sweden. He is also a professor at Lund University, from
where he holds a Ph.D. in information theory. In 1998, he joined Ericsson
Mobile Communication, where he worked on security solutions for
mobile phone platforms. His work greatly influenced the security solutions developed
for Ericsson Mobile Platforms. He also made major contributions to Bluetooth
security and platform security related patents. In 2005, he received the Ericsson
Inventors of the Year award and is currently working on trusted computing
technologies and the use of virtualization.
References
1. Ericsson Review, Setting the standard: methodology counters security
threats, January 2014, available at: http://www.ericsson.com/
news/140129-setting-the-standard-methodology-counters-security-threats_244099438_c
2. Stefan Berger, Ramón Cáceres, Kenneth A. Goldman, Ronald Perez, Reiner Sailer, Leendert van Doorn,
vTPM: Virtualizing the Trusted Platform Module, RC23879 (W0602-126) February 14, 2006, Computer
Science IBM Research Report, available at:
https://www.usenix.org/legacy/event/sec06/tech/full_papers/berger/berger.pdf
3. Tech Times, Cloud computing is the future but not if security problems persist, June 2014, available
at: http://www.techtimes.com/articles/8449/20140615/cloud-computing-is-the-future-but-not-ifsecurity-problems-persist.htm
4. Christian Huebner, Trusted Cloud computing with Intel TXT: The challenge, April 16, 2014, available at:
http://www.mirantis.com/blog/trusted-cloud-intel-txt-security-compliance/
5. Mudassar Aslam, Christian Gehrmann, Mats Bjorkman, Security and Trust Preserving VM Migrations in
Public Clouds, 2012 IEEE 11th International Conference on Trust, Security and Privacy in Computing and
Communications, June 25-27, 2012, available at:
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?reload=true&arnumber=6296062
6. Nicolae Paladi, Christian Gehrmann, Mudassar Aslam, Fredric Morenius, Trusted Launch of Virtual
Machine Instances in Public IaaS Environments, 15th Annual International Conference on Information
Security and Cryptology, 2013, available at:
http://soda.swedish-ict.se/5467/3/protocol.pdf
7. TrouSerS, the open source TCG Software Stack, I’ve taken ownership of my TPM under another OS...,
available at:
http://trousers.sourceforge.net/faq.html#1.7
8. Nicolae Paladi, Christian Gehrmann, Fredric Morenius, Domain-Based Storage Protection (DBSP)
in Public Infrastructure Clouds, 18th Nordic Conference, NordSec, October 18-21, 2013, available at:
http://link.springer.com/chapter/10.1007%2F978-3-642-41488-6_19#page-1
E R I C S S O N R E V I E W • 2/2014
Acknowledgements
The authors gratefully acknowledge the
colleagues who have contributed to
this article: Lal Chandran, Patrik Ekdahl,
András Méhes, Fredric Morenius,
Ari Pietikäinen, Christoph Schuba,
and Jukka Ylitalo
Re:view
27
25 years ago
The front cover of issue 4,
1989 depicted some of the
standardization organizations
of the time.
The associated article
discussed the role of
standardization in terms of
threats and opportunities,
concluding that the need
for it was more obvious
than ever. It noted that
Ericsson’s involvement in
standardization processes
Ericsson Review,
is necessary to be able to
issue 4, 1989.
influence the development of
technology.
50 years ago
Issue 4 in 1964 was dedicated to Ericsson’s new
telephone – the DIALOG. The design characteristics were
said to reflect the general
spirit of rationalization,
mechanization and functional
design permeating the era.
The telephone was starting to
play a central role in domestic
life and so not only its
functionality was addressed,
but also its appearance. Plugand-jack termination was
used to improve the mobility
of the device. Even at this
time, it was recognized that
Ericsson Review,
subscribers paying the same
issue 4, 1964.
amount for a given service
should enjoy the same quality
of transmission. The automatic regulation of transmission
level was an important step to reach this goal.
75 years ago
The fourth issue of 1939
carried an article on the neon
clock advertising department
store NK. Ericsson
constructed the illuminated
timepiece, claiming it to be the
biggest of its kind in Europe,
on Stockholm’s central
telephone tower. Despite a
subsequent fire in the tower,
the clock wasn’t damaged,
but was then moved to the NK Ericsson
Review,
store’s rooftop, where it still
issue 4, 1939.
stands today.
E R I C S S O N R E V I E W • 2/2014
High frequency small cell backhaul
28
Wireless backhaul in future
heterogeneous networks
Deploying a heterogeneous network by complementing a macro cell layer with a small
cell layer is an effective way to expand networks to handle traffic growth. For rollout to
be successful, however, relies on being able to provide all the additional small cells with
backhaul capability in a flexible and cost-efficient manner.
M I K A E L COL DR E Y, U L R I K A E NG S T RÖM , K E WA NG H E L M E R S S ON, MONA H A S H E M I , L A R S M A N HOL M , P ON T U S WA L L E N T I N
A number of proprietary wireless
small cell backhaul solutions
have been adapted to provide
carrier-grade performance in nonline-of-sight (NLOS) conditions.
These solutions typically operate
in both licensed and unlicensed
spectrum in the crowded sub6GHz frequency range. However,
to cope with predicted traffic load
increases, the need to exploit
additional spectrum at higher
microwave frequencies has been
identified.
This need led to Ericsson researching
how NLOS wireless backhaul could be
used at 28GHz. This research1 showed
how wireless small cell backhaul could
be implemented in an urban scenario
without a direct line-of-sight (LOS) path
between the deployed small cells and
the macro radio base station (RBS) providing backhaul connectivity1, 2. The
Ericsson research showed how pointto-point (PtP) microwave in licensed
spectrum could be used for small cell
NLOS backhaul, and2 showed that
point-to-multipoint (PtMP) could also
be used for the same purpose.
Building on this research, Ericsson
has investigated the impact on user
performance in a heterogeneous network of providing small cell backhaul
over a wireless link – by comparing it
with a system in which small cell backhaul is provided over (ideal) fiber. To do
this, a study was carried out using system simulations that captured the joint
impact of backhaul and access technologies on user performance. Two different
NLOS wireless backhaul technologies
were tested: a commercial high-end PtP
microwave backhaul and an LTE-based
PtMP concept – at 6GHz and 28GHz.
Both technologies were assumed to
operate in licensed microwave bands.
The results of the simulations show
that wireless backhaul technologies can
provide user performance on a comparable level to a fiber-based (ideal) solution. The results demonstrate that NLOS
backhaul deployed in licensed spectrum up to 30GHz is a future-proof technology that can manage high volumes
of traffic in heterogeneous networks.
BOX A Terms and abbreviations
EIRP
equivalent isotropic radiated power
EPC
Evolved Packet Core
EPS Evolved Packet System
IMT
International Mobile Telecommunications
ISD
inter-site distance
LOSline-of-sight
MIMO
multiple-input multiple-output
MTC
machine-type communication
E R I C S S O N R E V I E W • 2/2014
NLOSnon-line-of-sight
O&M
operations and maintenance
PtMPpoint-to-multipoint
PtPpoint-to-point
QAM
quadrature amplitude modulation
RAT
radio-access technology
UE
user equipment
WRC
World Radiocommunication Conference
Challenges created by small cells
Heterogeneous networks built by complementing a macro-cell layer with additional small cells in the RAN impose
new challenges on backhaul. For example, the best physical location for a small
cell often limits the option to use wired
backhaul. In urban areas, small cell
outdoor nodes are likely to be densely
deployed, mounted on lampposts and
building facades about three to six
meters above street level. If fiber exists
at the small cell site, it is the best option
for backhaul. But if fiber is not readily
available, deploying wireless backhaul
is both faster and more cost-effective.
Wireless backhaul is in itself nothing new, but small cell deployments
create new challenges for conventional
wireless backhaul, which was originally designed for LOS communication
from one macro site to another. In urban
environments and town centers, propagation paths between small cells and
macro sites are likely to be obstructed
by buildings, traffic signs and other
objects. Clear line-of-sight is highly
improbable. The number of users connected to each small cell might be just
a few, yet delivering superior and uniform user performance across the RAN
still requires a large number of small
cells. As a result, small cell backhaul
solutions need to be more cost-effective,
scalable, and simpler to install than traditional macro backhaul.
The dominant technology used in
backhaul networks today is based on
microwave – and predictions indicate
that this will continue to be the case. In
2019, microwave is expected to encompass about 50 percent of global backhaul
29
deployments3. The popularity of this
technology can be explained by the fact
that a microwave backhaul network
can be deployed quickly and in a flexible manner – two critical factors for
adoption.
The popularity of microwave has also
led to its extensive development over
the past few decades. For LOS deployments, microwave is capable of providing low cost, compact and easily
deployable backhaul capacity in the
order of several gigabits per second [4].
As mentioned, due to their placement
between street level and rooftop, a substantial portion of deployed small cells
will not have access to wired backhaul,
or have a clear LOS path to a macro site
with backhaul connectivity. These factors create a need for NLOS backhaul.
Solutions to the challenges posed
by NLOS conditions have already
been developed for microwave backhaul. Passive reflectors and repeaters
are sometimes used to propagate signals around obstacles in the communication path. However, this approach
is less desirable for cost-sensitive small
cell backhaul, as it increases the number of sites. Instead, providing singlehop wireless backhaul between a macro
site and a small cell site limits the number of sites needed, and is consequently
better suited to the small cell case. In
urban areas, daisy chaining can be used
to reach sites in difficult locations, and
this solution can also be used to advantage for small cell backhaul.
The propagation properties at lower
frequencies, below 6GHz, are well suited
for radio access. Consequently, modern
radio-access technologies (RATs) tend to
operate in licensed spectrum up to a few
gigahertz. Commercial microwave backhaul for macro sites operate at higher
frequencies – ranging from 6GHz to
70/80GHz. Operating small cell backhaul at these higher frequencies allows
spectrum in the lower frequency bands
to be used by radio access, which leads to
­better spectrum utilization overall.
Joint access and backhaul
In 5G networks, it is likely that access
and backhaul will, to a large extent, converge: in some deployments, the same
wireless technology can be used effectively for both. This convergence may
lead to more efficient use of spectrum
FIGURE 1 Example of LTE-based PtMP backhaul system architecture
Macro RBS
and hub
3GPP
core
User
Small RBS
and client
resources, as they can be shared dynamically between access and backhaul5. For
other deployments, a complementary
and more optimized backhaul solution
might be the preferred choice to support 5G features, such as guaranteed
low latency at an extremely high reliability for mission critical MTC, as this
is more backhaul critical.
Another more high-level benefit of
convergence is the ability to use the
same operations and maintenance
(O&M) system for access and backhaul,
which can both improve overall system
performance and simplify system management. For example, a common network management that can combine
KPIs from the entire network can make
optimized decisions and take effective
action to improve overall performance.
Such KPIs include data rates, latencies, and traffic loads experienced by
the various nodes in a heterogeneous
network; including macro cells, small
cells, and backhaul. If not impossible,
such network performance optimization becomes extremely challenging if
the KPIs are inaccessible and the nodes
are uncoordinated. A common network
management system is, therefore, an
enabler for efficient operation of a heterogeneous network.
Irrespective of convergence, the costeffectiveness of backhaul connections
becomes increasingly important in
deployments that include large numbers of small cells. In general, deployments that have less hardware and
simplified installation procedures
are more cost-effective. So, as PtMP
User
backhaul connections simplify deployment, applying this technology is one
way to reduce costs.
In the present study, a system level
approach was used to evaluate the joint
effect of converged access and backhaul.
A complete heterogeneous LTE RAN
deployed in a dense urban scenario was
simulated encompassing macro cells,
small cells, small cell backhaul, users,
traffic models, propagation, interference, and scheduling effects. Using such
an advanced simulation environment
makes it possible to evaluate overall system and user performance for different
small cell backhaul scenarios in a way
that captures the joint impact of access
and backhaul.
Backhaul technologies
for small cells
The various technologies that exist
for wireless backhaul can be classified
into two main solution groups: PtP and
PtMP. A PtP solution uses dedicated
radios and narrow-beam antennas to
provide backhaul between two nodes.
In a PtMP solution, one node provides
backhaul to several other nodes by sharing its antenna and radio resources. As
illustrated in Figure 1, the nodes in a
PtMP scenario are referred to as hub
and client, where the hub is typically
colocated with a macro site (that has
backhaul connectivity) and the client is
colocated with a small cell site.
Spectrum
Irrespective of the technology deployed,
user performance is directly
E R I C S S O N R E V I E W • 2/2014
High frequency small cell backhaul
30
FIGURE 2 NLOS wireless backhaul client/hub – urban deployment
Hub
Hub
Client
Client
related to optimal use of spectrum.
The 2015 World Radiocommunication
Conference (WRC-15) will focus on the
future allocation of additional spectrum below 6.5GHz for radio access.
Looking at current spectrum allocation,
these frequencies are crowded, which
means that the potential for more backhaul bandwidth in licensed spectrum
is greater for frequencies above this.
Backhaul based on Wi-Fi and LTE are
just two of the current technologies
operating below 6GHz. Wi-Fi typically
operates in unlicensed spectrum and is
therefore prone to interference while,
for example, LTE relaying exploits
licensed IMT spectrum for both backhaul and access.
Using unlicensed frequency bands
might be a tempting option to reduce
cost, but this approach can result in
unpredictable interference issues that
make it difficult to guarantee QoS. The
potential risk associated with unlicensed use of the 60GHz band is, however, lower than the risk associated with
the popular 2.4GHz and 5GHz bands.
This is due to very high atmospheric
attenuation caused by the resonance of
oxygen molecules around 60GHz and
the possibility to use compact antennas with narrow beams – which reduce
interference effectively.
The conventional and spectrum-efficient licensing policy for PtP microwave
backhaul works on an individual linkby-link licensing basis6. However, when
it comes to rolling out small cell backhaul, simplicity, multipath interference
E R I C S S O N R E V I E W • 2/2014
issues, and cost are of such importance
that other policies for licensing should
be considered.
Light licensing and block licensing
are two possible alternatives. In the
light licensing case, license application
is a simple and automated process that
involves only a nominal registration
cost. This approach can be used in scenarios where interference is not a major
concern or can be mitigated by technical means6. It has become popular to use
light licensing to encourage the uptake
of PtP E-band links. If properly deployed,
these communication links do not interfere with each other due to high atmospheric absorption and narrow beam
widths.
In block or area licensing, the licensee
has the freedom to deploy a radio emitter within a given frequency block and
geographic area as long as the radio fulfills some basic requirements, such as
respecting the maximum equivalent
isotropic radiated power (EIRP). In this
case, the licensee is responsible for managing co-channel interference between
different transmissions and making it
suitable for managing PtMP backhaul
and radio access systems7.
Being able to exploit the spectrum
potential offered by higher frequency
bands from 10GHz to 100GHz is part
of ongoing research for 5G5,8. The high
propagation losses that are associated with high-frequency millimeter
waves typically limit the applicability
of such high frequency bands to shortrange links. These losses can be partly
compensated for with more advanced
antenna systems using beamforming.
However, this makes mobility at high
speeds (such as in cars and on highspeed trains) more challenging, as
beams would need to be adapted more
or less continuously.
Wireless backhauling of fixed nodes
is less of a challenge, as alignment or
beam pointing is more straightforward
when nodes are situated in predefined
fixed locations than when they are constantly moving – and so the application
of higher frequencies is simpler.
Capacity and availability
Backhaul capacity is often dimensioned to support the peak capacity of
the macro cell9. However, in practice,
the trade-off between cost and the need
for capacity usually results in a more
practical level for backhaul capacity
being set. This level should, at a minimum, support expected busy-hour traffic, with some margin to account for
statistical variation and future growth.
Dimensioning in this way makes sense
when it comes to cost-sensitive small
cell backhaul. However, it is recognized
that different operators – to align with
their business strategy – are likely to use
different approaches for capacity provisioning of small cell backhaul.
Today’s minimum bitrate targets
for backhauling 3GPP LTE small cells
is somewhere in the region of 50Mbps
for radio access using 20MHz of spectrum. To support current peak rate
demands, however, 150Mbps or more is
desirable9. These targets for minimum
and peak bitrates are likely to increase
further over the next few years as traffic volumes continue to rise, and additional spectra and new features for radio
access become available. In addition,
small cell access points may not only be
required to support multiple 3GPP technologies (such as HSPA and LTE) but may
also include Wi-Fi, which will further
increase the need for backhaul capacity.
Availability requirements may differ between small cell and macro cell
backhaul, depending on the deployment scenario. The availability requirement for macro backhaul can be as high
as 99.999 percent (which corresponds to
a maximum of five minutes of outage
per year). For small cell backhaul, such
high availability requirements may not
31
be necessary. If the small cell is deployed
to boost data rates or capacity in an area
with existing macro coverage, the backhaul requirements could be relaxed significantly to, for example, 99-99.9 percent
(which corresponds to anywhere from 12
hours up to several days of outage per
year)8.
From a user perspective, the performance of an individual backhaul link is
less relevant. What matters is the overall performance of the combined backhaul and access links. If the access link
at a given time and place provides a certain level of service, the corresponding backhaul link does not need to be
significantly better. Hence, the access
and backhaul links could be jointly
optimized. To reflect this in the present study, the joint effect of access and
backhaul on user performance was evaluated, using an all LTE-based backhaul
concept operating at higher frequencies that is more integrated with the
LTE access than conventional wireless
backhaul.
Antennas
Maximum antenna gain is given by the
antenna size in relation to the wavelength of the frequency used. As a result,
antennas that are smaller in size than
antennas with the same antenna gain
at lower frequencies can be deployed at
higher frequencies. If aligned correctly,
a compact high-gain antenna can compensate for the increased path loss that
is usually associated with higher frequencies and NLOS conditions.
A PtP system uses high-gain antennas at both ends of a link, while a PtMP
system uses a wide-beam antenna at the
hub site and a directive antenna at the
client site.
More advanced antenna solutions at
the hub site, such as steerable or fixed
narrow multi-beam systems, can be
deployed, but such solutions will probably not be cost-effective for some time.
Carrying out manual antenna alignment with narrow beam widths in
NLOS conditions may sound like a difficult task, but it can be a surprisingly simple procedure, even at 28GHz1. However,
as correct alignment is important, especially at higher frequencies, it may be
a good idea to deploy a client antenna
that has automatic beam-steering capabilities, so that it can simply align itself
to the best signal path. Beam steering
can be implemented using mechanical
methods, antenna arrays or a combination of the two.
LTE-based backhaul concept
To address the issue of providing backhaul in heterogeneous networks, a new
concept is being researched based on
the adaptation of LTE technology for
small cell backhaul at high microwave
frequencies – evaluated at 6GHz and
28GHz.
This concept reuses the LTE physical
layer but applied at a higher frequency
band – up to 30GHz. As LTE physical-layer numerology was originally
designed to operate with a carrier frequency of around 2GHz, operation in
higher bands requires some modification of the original concept. But if top-ofthe-line hardware is in place, the need to
change the numerology (by increasing
the subcarrier spacing, for example) for
frequencies below 30GHz in a backhaul
context is small. However, to reduce
hardware costs, numerology may need
to be adjusted to match higher microwave frequencies. This concept is part
of 5G ­radio access research5.
With a 3GPP LTE-based PtMP solution, backhaul links can inherit 3GPP
functionality already developed for
LTE access, as well as features that will
be implemented in the future, such as
carrier aggregation, reduced latency,
advanced schemes for beamforming,
MIMO, interference cancellation and
radio resource scheduling. When backhaul and access links are converged,
operational efficiency can be increased,
as the overhead created by managing
different technologies is reduced. For
example, the control and management
architecture as defined by the 3GPP
Evolved Packet System (EPS) can be used
by both systems.
An example system architecture for
LTE-based PtMP backhaul is illustrated
in Figure 1. The basic principles of this
architecture include interfaces, protocols, the reuse of 3GPP logical nodes,
EPS bearer concept, as well as security
solutions.
As Figure 1 illustrates, the small RBS
is connected to a client. The client provides the wireless backhaul IP-based
transport to the core network, which
in turn provides functions like bearer
management, QoS enforcement and
authentication. The client terminates
the LTE radio interface and implements UE functions such as cell search,
measurement reporting, and radio
transmission and reception. The hub
implements the eNodeB side of the LTE
radio interface. In this example, both
the hubs and the clients are controlled
by a 3GPP-based EPC network – which
can be a core network dedicated to backhaul, or a core network shared between
the small RBS and the access links.
While there are similarities between
an all-LTE network (backhaul plus
access) and the LTE relay solution developed in 3GPP (which also provides backhaul based on an LTE radio interface),
there are two main differences between
them. First, LTE backhaul has been modeled as a transport network. As such, it
is access-agnostic and can be used with
any access link technology. LTE relay on
the other hand has been designed to use
LTE link technology for both backhaul
and access. The second difference is that
LTE backhaul links and LTE access links
typically use separate radio resources
(separated in terms of frequency bands),
while the (in-band) LTE relay solution
shares radio resources between the
backhaul and access links.
In summary, an LTE-based PtMP
backhaul provides several benefits compared with other alternatives:
reuse of functionality – inherent
multiple access (PtMP), architecture,
protocol structure, physical layer,
procedures, and security mechanisms
are just some examples of functionality
already developed in 3GPP;
quick launch of new features – by
reusing existing (and future) LTE
developments, new features can also be
rapidly deployed;
use of the same ecosystem – one system
for both backhaul and access links can
simplify O&M for operators and increase
operational efficiency;
support for multi-RAT access links –
compared with LTE relaying solutions,
any RAT can be used on the access link;
joint backhaul-access link optimization
– added value can be achieved through
dynamic optimization and operation of
access and backhaul targeting user
performance. A high level of integration
and potentially shared hardware are
E R I C S S O N R E V I E W • 2/2014
High frequency small cell backhaul
32
FIGURE 3
European deployment scenario
Throughput
Path loss
Mb/s
dB
120.0
110.0
100.0
90.0
80.0
70.0
60.0
50.0
40.0
30.0
20.0
10.0
0.0
130.0
120.0
110.0
100.0
90.0
80.0
70.0
60.0
50.0
40.0
30.0
20.0
Small RBS and client site
US deployment scenario
Throughput
Path loss
Mb/s
dB
120.0
110.0
100.0
90.0
80.0
70.0
60.0
50.0
40.0
30.0
20.0
10.0
0.0
130.0
120.0
110.0
100.0
90.0
80.0
70.0
60.0
50.0
40.0
30.0
20.0
Macro RBS and hub site
Small RBS and client site
other potential benefits of converged
links; and
automated deployment – installation
procedures similar to those used to set
up a small RBS (which today is
automatic) and can also be used to
install the backhaul client.
Evaluation scenarios
In this study, heterogenous networks
were simulated using macro and small
cells for radio access and hubs and clients for wireless backhaul deployed in
E R I C S S O N R E V I E W • 2/2014
building heights are assumed to be
homogenous, ranging from 5m to 40m;
no high-rises;
few open areas;
19 macro/hub sites with an average ISD
of 400m; and
76 small RBS/client sites.
The US city environment is more challenging, assuming that:
a downtown area exists with high-rises
as well as surrounding low buildings,
with open spaces in between;
building heights range from 4m to
288m;
19 macro/hub sites with an average ISD
of 700m; and
114 small RBS/client sites.
Macro RBS and hub site
FIGURE 4
PtMP concept (described in this article). Figure 2 illustrates the simulation
scenario, showing two hubs providing
wireless backhaul to two clients in an
urban environment.
Some assumptions were made about
the nature of the virtual cities. For the
European city:
two virtual cities. These cities aimed to
represent a typical European scenario
with a dense macro deployment and
a typical US scenario with downtown
high rises and a sparse macro deployment with a greater number of small
cells per macro.
The macro RBSs and backhaul hubs
were colocated at the same site, as were
the small RBSs and clients. The clients
were located above street level and backhauled wirelessly to a serving hub using
either PtP microwave or the LTE-based
Figures 3 and 4 illustrate a portion
of the deployments for the virtual
European and US cities. The left side
of each figure shows the results of the
macro-only network, and the right side
shows the results of a combined macro
and small cell deployment that uses
LTE-based PtMP backhaul at 28GHz.
The colors of the cells indicate average
user throughput, according to the scale
on the left. The line between a hub and
a client shows the strongest propagation path, and the color of the line indicates its path loss. The improvement in
throughput, illustrated by the amount
of green in the illustrations, due to
offloading of the macro in the small cell
deployment is considerable. The simulated served traffic levels in the network
are 20GB/month/user in the European
scenario and 6GB/month/user in the US
scenario.
For LTE access, the simulated carrier frequencies were set to 2.1GHz in
the European scenario and 700MHz in
the US scenario. The access bandwidth
was 20MHz in both cases, which corresponds to a peak rate of 108Mbps using
2x2 MIMO. The macro RBS output
power was assumed to be 2x30W and
the small RBS output power to be 2x5W.
High-gain backhaul antennas were
used to compensate for the greater NLOS
33
path loss at higher microwave frequencies. In the PtP evaluations, mechanically steerable high-gain antennas were
used at both the hub and client sites,
while for PtMP evaluations, the hub was
implemented using fixed sector-covering antennas. Antenna parameters and
output power of hub and client for the
different backhaul systems and carrier
frequencies are summarized in Table 1.
For PtMP, 20MHz of bandwidth at two
frequencies were evaluated – 6GHz and
28GHz – while only 28GHz was considered in the PtP case. The LTE-based PtMP
used fixed output power in the downlink, while PtP used adaptive power
control.
Methodology
User performance including wireless
backhaul was evaluated in a static system simulator. In the simulator, LTE
access was based on LTE Rel-8 with 2x2
MIMO and 64QAM in the downlink,
which corresponds to a downlink peak
rate of 108Mbps when using 20MHz of
access bandwidth. The wireless backhaul, including LTE-based PtMP and
commercial PtP microwave, were also
simulated using 20MHz bandwidth. In
one simulated case, 40MHz was also
used for the LTE-based PtMP backhaul
for the more challenging US scenario, to
illustrate the use of the LTE feature carrier aggregation on the backhaul.
User-generated traffic for both simulation scenarios was split on an 80/20
basis – 80 percent generated by indoor
users and 20 percent by people outdoors.
Indoor users were evenly distributed
among the floors of the buildings, and
traffic load was measured in terms of
data traffic consumed by one user in one
month. For each scenario and deployment, as traffic load increased, the traffic served by the system increased until
the system reached its capacity limit.
This limit depends on the scenario and
the deployment, including the number
of macro RBSs and small RBSs deployed.
To put some perspective on the traffic
load, 2014 levels for actual mobile traffic
are in the region of 1.5-2GB /user/month
in Europe and the US. Mobile data traffic is expected to grow globally by 45
percent annually 2013-2019, so by the
end of 2019, mobile traffic will be somewhere around 10GB /user/month3.
User throughput is given by the size
of a data packet and the total transmission time of the packet. The transmission time takes into account any
delay due to resource sharing: multiple users accessing the same radio
resources. Each user is served either by
a macro or by a small RBS. For those
served by a macro, only resource sharing on the access side has an impact on
throughput. For users served by small
RBSs, aside from the resource-sharing
delay on the access side, there is also a
resource-sharing delay associated with
the wireless backhaul. Resource sharing in the backhaul results from either
multiple users connected to the same
small RBS – which means they share its
backhaul connection – or from users
connected to different small RBSs that
share a common backhaul connection
in a PtMP situation. As each PtP backhaul link has an individual (not shared)
backhaul resource, PtP backhaul is only
shared by users connected to the same
small RBS. However, the PtMP backhaul
may be shared by users connected to different small RBSs that are connected to
the same hub sector. Hence for small
RBS users, user performance depends
not only on the access but also on the
type of backhaul that carries the small
RBS traffic.
Wrap up
European city scenario
Figure 5 shows user throughput (in
the downlink) against served traffic
for the European scenario. The curves
represent the macro-only network (blue
curves) as well as heterogeneous networks with three different small cell
backhaul technologies (yellow, red and
purple curves), according to:
yellow – PtP microwave at 28GHz with
20MHz bandwidth;
red – LTE-based PtMP at 28GHz with
20MHz bandwidth; and
purple – LTE-based PtMP at 6GHz with
20MHz bandwidth.
The reference performance levels for
fiber backhaul (green curve) are also
shown. The 10th percentile represents
the 10 percent worst case rates experienced by users, the 50th represents the
median, and the 90th percentile represents the top 10 percent downlink performance rates.
The immediate conclusion from this
is that small cell deployment can radically improve user throughput, especially at high traffic levels where the
macro-only network cannot meet the
demand.
When looking at the served traffic levels, the network has a very good
macro deployment, as it alone can serve
10GB/user/month while maintaining a
10th percentile downlink user throughput of about 10Mbps. By deploying small
cells, the corresponding user throughput is increased to 30Mbps, or the 10th
percentile at 10Mbps is maintained,
while the network serves as much as
23GB/user/month.
Table 1: Antenna parameters and output powers for the different backhaul systems
Node type
Frequency
[GHz]
Antenna
type
Azimuth
HPBW1
[degrees]
Elevation
HPBW1
[degrees]
Max. gain
[dBi]
Aperture
size
Max. power
[dBm]
28
Sector
65°
5°
20
1.5 x 12.5 [cm2]
23
6
Sector
65°
5°
20
6.5 x 54 [cm2]
23
28
Parabolic
reflector
3°
3°
34
Diameter = 20 [cm]
23
6
Patch array
14°
14°
22
20 x 20 [cm2]
23
28
Parabolic
reflector
3°
3°
34
Diameter = 20 [cm]
23
PtMP hub
PtMP client
PtP client
and hub
1
half power beam width
E R I C S S O N R E V I E W • 2/2014
High frequency small cell backhaul
34
As expected, the choice of small
cell backhaul has almost no impact on
the worst case 10th percentile, as these
users are more limited by the access
network than by the backhaul. Small
backhaul limitations only occur for the
median (50th percentile) and best (90th
percentile) users connected via PtMP
backhaul – observed by small penalties
compared with fiber. The PtP backhaul
shows close-to-fiber performance for all
users and served traffic levels. It is also
noticeable that all backhaul options can
cope with the user peak rates (108Mbps)
achieved at lower loads (90th percentile
and below 10GB/month/user).
The variation in performance
between PtP and PtMP wireless backhaul is due to two primary differences
in these systems. Firstly, two different
antenna systems are used, where PtMP
has wide-beam sector antennas at the
hub, while PtP has directive high-gain
antennas at both ends of each link. The
PtMP sector antenna has a much lower
antenna gain than the narrow beam
PtP antenna – 14dB lower, as shown
in Table 1. Secondly, there is less sharing of resources in the PtP backhaul,
where each client has its own dedicated
resource, while the PtMP system may
also share its resources over multiple clients. In the simulated PtMP case, a hub
has three sectors and each sector may
serve one to five clients depending on
the traffic load in that sector.
Finally, the performance levels of the
PtMP backhaul operating at 6GHz and
28GHz are almost identical. Both systems have identical antenna gain and
beamwidth at the hub, while the 6GHz
system has 12dB lower antenna gain
and wider beamwidth at the client. On
the negative side, a lower antenna gain
results in worse system gain and a wider
beamwidth is more prone to interference. However, on the positive side, the
6GHz system experiences less path loss,
which compensates the negative side.
FIGURE 5 European scenario
User throughput (Mbps)
120
Macro
Fiber
PtP microwave; 28GHz, 20MHz
LTE-based PtMP; 28GHz, 20MHz
LTE-based PtMP; 6GHz, 20MHz
100
50th percentile
90th percentile
80
60
40
10th percentile
20
0
0
5
10
15
20
25
30
35
40
Served traffic (GB/month/user)
E R I C S S O N R E V I E W • 2/2014
US city scenario
Figure 6 presents the downlink user
throughput against served traffic in the
US city. The network capacity in this
scenario is limited by the macro network since the macro network is much
sparser than the European city. This
is observed in the much lower served
traffic values and the poor macro-only
performance. Deploying small cells
improves the network performance
substantially.
Also in this scenario, worst case user
perfomance (10th percentile) is limited
by access and not by backhaul, so the
choice of backhaul has no impact on
worst case user throughput. But when
looking at best case user performance
(90th percentile), there is a clearer backhaul limitation when using PtMP backhaul with 20MHz bandwidth at higher
served traffic levels. A remedy for
improving PtMP performance for high
performance users is to apply the LTE
feature carrier aggregation in the LTEbased PtMP backhaul. Figure 6 shows
the result when a 40MHz bandwidth
is applied to the backhaul at 28GHz
and the user performance is improved
and PtMP with carrier aggregation is
on a par with PtP microwave and fiber.
Thanks to reduced resource sharing and
high-gain antennas at both ends, the PtP
backhaul also shows close-to-fiber performance for all users and served traffic
levels in this scenario.
When comparing PtMP at 6GHz
to 28GHz, some degradation for high
throughput users is observed in the
90th percentile at high traffic levels
in Figure 6. This is due to the different antenna characteristics, where the
antenna gain at 28GHz is 12dB higher at
the client side than it is at 6GHz and the
wider client antenna beam at 6GHz has
less spatial filtering of interference compared with the 28GHz client antenna.
Summary
Deploying small cells provides a means
for handling future traffic growth and
enables a substantial improvement in
network performance. It is therefore of
great importance to enable small cell
deployments by providing cost-effective backhaul. The study carried out
addresses some of the challenges created by small cell backhaul. By using system simulations that capture the joint
35
effect of access and backhaul, it has
been shown that NLOS microwave backhaul in licensed spectrum up to 30GHz
is a viable solution for dense small cell
deployments in urban environments.
A novel LTE-based NLOS PtMP backhaul concept operating at high microwave frequencies, up to 30GHz, has also
been evaluated. This concept is a potential step toward using LTE at higher frequencies and converging access and
backhaul networks, which is also foreseen in 5G networks.
System simulations for two different
deployment scenarios show that degradation in user performance is minimal
when wireless backhaul is compared
with (ideal) fiber backhaul – for lower
to medium throughput users. For high
throughput users, the performance of
the LTE-based NLOS PtMP backhaul concept is not as good as the PtP microwave
backhaul – which shows close-to-fiber
performance for all users and served
traffic levels due to greater numbers of
radio and antenna resources. The LTEbased NLOS PtMP backhaul was evaluated both at 6GHz and 28GHz, and
28GHz works just as well or even better
than 6GHz.
In the more challenging US deployment scenario, the performance degradation with LTE-based PtMP was
rectified by applying larger bandwidth
in the microwave backhaul by using carrier aggregation, which is inherent in
LTE, bringing it up to par with NLOS PtP
and fiber backhaul.
FIGURE 6 US scenario
User throughput (Mbps)
120
Macro
Fiber
PtP microwave; 28GHz
LTE-based PtMP; 28GHz, 20MHz
LTE-based PtMP; 6GHz, 20MHz
LTE-based PtMP; 28GHz, 40MHz
100
80
60
90th percentile
90th percentile
40
50th percentile
20
10th percentile
50th percentile
10th percentile
0
2
4
6
8
10
12
14
Served traffic (GB/month/user)
E R I C S S O N R E V I E W • 2/2014
High frequency small cell backhaul
36
Mikael Coldrey
Mona Hashemi
holds an M.Sc. in applied physics and electrical
engineering from Linköping University, Sweden, and a Ph.D.
degree in electrical engineering from Chalmers University of
Technology, Gothenburg, Sweden. He joined the Radio
Access Technologies department within Ericsson Research in 2006,
where he is a senior researcher. He has been working with both 4G and
5G research. His main research interests are in the areas of advanced
antenna systems, models, algorithms, and millimeter wave
communications for both radio access and wireless backhaul systems.
Since 2012, he has also been an adjunct associate professor at Chalmers
University of Technology.
joined Ericsson Research in 2010 after completing her
M.Sc. in wireless and photonics engineering at Chalmers
University of Technology, Gothenburg, Sweden the same
year. She holds an experienced researcher position at
Ericsson Research, and has been involved in a variety of projects, such
as the NLOS wireless backhaul, and the EARTH project founded by the
Seventh Framework Programme (FP7) of the European Commission.
Currently, she is working on standardization and concept evaluation
for LTE.
Ulrika Engström
Lars Manholm
received his M.Sc. in electrical engineering, and his
Lic. Eng. in electromagnetics from Chalmers University of
Technology, Gothenburg, Sweden in 1994 and 1998,
respectively. He joined Ericsson as an antenna designer in
1998 and moved to Ericsson Research in 2003. He is currently working
as a senior researcher focusing on antennas for millimeter wave and
higher microwave frequencies.
Ke Wang Helmersson
joined Ericsson Research in 1995 and is currently
working in the Wireless Access Networks department at
Ericsson Research in Linköping, Sweden, where she is a
senior researcher in RRM and system-level simulations, as
well as performance evaluations. She has been involved in research
and development efforts for EDGE, HSPA and LTE and wireless
backhaul technologies. She is currently working on future wireless
industrial applications in the 5G program. She holds a Ph.D. in electrical
engineering from Linköping University, Sweden.
received a Ph.D. in physics from Chalmers University of
Technology, Gothenburg, Sweden, in 1999, and an M.Sc. in
physics and engineering physics, also from Chalmers in
1994. She joined the antenna research group at Ericsson
Research in Gothenburg, Sweden in 1999. Her main research focus is
antenna systems, targeting wireless backhaul challenges for small
cells, LTE and 5G. She has had a variety of roles in, for example,
Ericsson’s testbed development and system evaluations, including
serving as project manager of several successful research projects
within Ericsson Research. She is currently driving studies within the 5G
program at Ericsson.
Pontus Wallentin
is a master researcher at Ericsson Research, wireless
access networks. He joined Ericsson in 1988 working with
GSM and TDMA system design. Since joining Ericsson
Research in 1996, he has focused on concept development
and 3GPP standardization of 3G WCDMA/HSPA and LTE. He holds an
M.Sc. in electrical engineering from Linköping University, Sweden.
References
1. Ericsson Review, 2013, Non-line-of-sight microwave backhaul for small cells, available at:
http://www.ericsson.com/res/thecompany/docs/publications/ericsson_review/2013/er-nlos-microwave-backhaul.pdf
2. IEEE Communications Magazine, 2013, Non-line-of-sight small cell backhauling using microwave technology, available at:
http://dx.doi.org/10.1109/MCOM.2013.6588654
3. Ericsson Mobility Report, June 2014, available at: http://www.ericsson.com/res/docs/2014/ericsson-mobility-report-june-2014.pdf
4. Ericsson Review, 2011, Microwave capacity evolution, available at: http://www.ericsson.com/res/docs/review/Microwave-Capacity-Evolution.pdf
5. Ericsson Review, 2014, 5G radio access, available at:
http://www.ericsson.com/res/thecompany/docs/publications/ericsson_review/2014/er-5g-radio-access.pdf
6. Electronic Communications Committee (ECC), Report, Light licensing, license exempt and commons, Report 132, 2009, available at:
http://www.erodocdb.dk/Docs/doc98/official/pdf/ECCRep132.pdf
7. Electronic Communications Committee (ECC), Report, Fixed service in Europe – current use and future trends post, Report 173, 2012, available at:
http://www.erodocdb.dk/Docs/doc98/official/pdf/ECCRep173.PDF
8. IEEE Access, vol. 1, May 2013, Millimeter wave mobile communications for 5G cellular: It will work!, available at:
http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=6515173
9. NGMN Alliance, White Paper, 2012, Small Cell Backhaul Requirements, available at:
http://www.ngmn.org/uploads/media/NGMN_Whitepaper_Small_Cell_Backhaul_Requirements.pdf
E R I C S S O N R E V I E W • 2/2014
Re:view
37
25 years ago
The cover of issue 3, 1989 shows a batch of wafers
being fed into an LPCVD furnace for coating with silicon
nitride. Semiconductor
technology in general, and
MOS technology in
particular, was
developing so that more and
more circuit elements could
be made on a single chip.
The article describes the
state of MOS technology at
the time and some
development trends. At the
time, Ericsson Components
AB manufactured electronic
Ericsson Review,
components, including
issue 3, 1989.
printed circuit boards and
fiber optics.
50 years ago
The cover of issue 3,
1964 shows part of a rack in
an automatic code switch
exchange. The lead article
stressed the design
elements of exchange
technology and how
Ericsson developed a
selector designed to meet
substantially reduced space
requirements and overall
capital investment. This was
a design move away from
the crossbar system that
had been in use in Ericsson’s
systems since the 1920s.
Ericsson Review,
issue 3, 1964.
75 years ago
The third issue of 1939
carried an article on the leading
telephone cities of the world.
Cities were ranked in terms of
telephone density for the
period 1929-1937. Washington
DC topped the list, with San
Francisco a close second. With
Stockholm in third position it
outranked all other European
cities by a long way with
American cities occupying all
other leading positions. Today, Ericsson Review,
Ericsson’s City Index shows a issue 3, 1939.
much wider and more even
spread across the globe.
E R I C S S O N R E V I E W • 2/2014
Indoor made simple
38
Connecting the dots: small cells
shape up for high-performance
indoor radio
In 2012, the global consumption of mobile data traffic in a month amounted to 1.1 exabytes. This figure
is set to rise to 20 exabytes by 2019, corresponding to a CAGR of 45 percent1. Today, this traffic is split
70/30 with the larger proportion consumed indoors; a level that is not expected to decrease. Adapting
networks to support such a rapid rise in traffic demand will require massive deployments of targeted
indoor small cell solutions, complemented by denser outdoor deployments.
CH E NGUA NG LU, M IGU E L BE RG, E L M A R T ROJ E R , PE R-E R I K E R I K S SON, K I M L A R AQU I , OL L E V. T I DBL A D, A N D H E N R I K A L M E I DA
How do you design a small radio
to fit the interiors of large spaces,
yet powerful enough to meet
future requirements for indoor
radio capacity? This was the
question we asked ourselves
when we began to develop a
solution to provide high-capacity
radio for indoor environments.
radio. This article presents how we overcame the challenges.
Managing mobile data traffic volumes is already a challenge in many
markets, and as traffic trends continue
to rise, the need to efficiently manage
indoor traffic becomes more significant.
Some of the factors contributing to the
challenge of data traffic are:
What we wanted was a solution that
could provide high-performance connectivity, in the increasingly demanding indoor radio environment. We
wanted the installation process to be
simple and to reuse existing building infrastructure. We needed to find
an efficient way to deliver power and
a design that integrates well with outdoor solutions.
The result, the Ericsson Radio Dot
System (RDS), is a novel indoor small cell
solution with a flexible radio architecture for providing high-capacity indoor
Meeting the requirement for more
indoor capacity calls for a combination of macro network extension and
new energy-efficient building standards
– resulting in higher attenuation in outer
walls and windows;
global urbanization development –
today, 54 percent of the world’s
population live and work in dense city
environments, a figure that is forecast to
rise to 66 percent by 20502; and
the gradual consumption shift from
laptops to smartphones3 boosted by
network enablers, application
adaptations, and device evolution.
BOX A Terms and abbreviations
ACLR
CAGR
CPRI
DAS
DU
FDD
IF
IRU
MIMO
O&M
PCC
adjacent channel leakage ratio
compound annual growth rate
Common Public Radio Interface
distributed antenna system
digital unit
frequency division duplexing
intermediate frequency
indoor radio unit
multiple-input, multiple-output
operations and management
primary component carrier
E R I C S S O N R E V I E W • 2/2014
PoE
RDS
RF
RRU RU
SCC
SDMA
SINR
TCO
TDD
UE
Power over Ethernet
Radio Dot System
radio frequency
remote radio unit
radio unit
secondary component carrier
spatial division multiple access
signal-to-interference-plus-noise ratio
total cost of ownership
time division duplexing
user equipment
densification, together with specific targeted indoor small cell solutions.
To handle peak rates, high capacity
small cells require the same level of
backhauling capabilities and baseband
processing as larger cells. However,
when compared with larger cells, the
cost of backhauling and other resources
(such as baseband processing capability) for small cells typically needs to be
balanced against the fewer numbers
of users served. So, the ability to simplify backhauling and provide a means
to support shared baseband and higher
layer processing across many small cells
becomes critical.
Femtocell-like solutions, with baseband and cell definition at the antenna
point, were thought to be candidates for
indoor capacity needs. Unfortunately,
these types of nodes only work in practice for small deployments, because
radio coordination and cell planning
quickly become unmanageable as the
number of cells increases. For medium
to large buildings, venues and arenas,
macro cell features like coordination,
seamless mobility and interference
management are needed. Supporting
these features points us in the direction
of concepts like main-remote and fronthauling, and solutions that use common baseband processing for remotely
deployed small cell radio heads.
For small cell indoor scenarios, the
preferred transmission medium is
largely dictated by economies of scale.
For example, the ability to use the same
type of cabling and building practices
39
FIGURE 1 Reference architecture – distributed antenna system
RBS with regular
radio units
Attenuator bank
DAS head-end
DAS remote unit
Base station
integration unit
DAS antennas
Radio
distribution units
Coaxial
Fiber
Optical
distribution units
Extended power,
backup and cooling
as those that the IT industry uses for
Ethernet services would be advantageous for any solution. Twisted-pair
copper LAN cables are particularly
attractive, as they tend to be deployed
abundantly within enterprises and are
widely supported by the IT community.
Installing these cables is a relatively simple process, as it does not require specially trained staff or expensive tools.
And in addition, the whole IT ecosystem for LAN cables can be leveraged –
from installation and support staff, to
established installation and maintenance practices, as well as technologies
for fault localization and diagnosis.
Making use of LAN cables is one important characteristic of the Ericsson RDS,
which also benefits from being able
to reuse existing tools developed for
fault localization, diagnosis and copper
cabling. Using copper cables to connect
radio equipment has the additional benefit of remote powering – power is fed
over the same medium as the communications signals. This reduces the complexity and cost of installation, as there
is no longer any need to arrange for local
power, which can be a costly process.
Remote powering from a central location makes it much easier to provide
backup power at the central location,
thereby increasing reliability.
Remote power
Fiber
termination unit
Fiber DAS (parallel)
The major challenge of a traditional fronthauling solution over LAN
cables is meeting the requirements for
latency and its variation, as well as for
high capacity and reach. With the current limitations of the CPRI protocol4, it
would not be possible to apply a mainremote (digital unit (DU)-radio unit
(RU) split) concept over longer distances
using copper cables. As discussed later
in this article, there are additional reasons – like power efficiency and small
cell complexity – for not pursuing CPRI
as it is currently specified.
Our mindset during the conceptualization of the RDS was one of rethinking the ecosystem around how to secure
radio access capacity for indoor environments, taking costs into consideration,
as well as simplicity of installation and
operations, power feeding and the existing indoor infrastructure. We wanted
to create a solution that would fully
unleash the capabilities of existing and
future radio-access solutions and all of
their features5.
Our starting point was to take a view
of the indoor small cell as an extension
to and an enhancement of the macro cellular network. We revisited the RU architecture in such a way that deployed radio
heads would be connected to the rest of
the network via LAN cables, while still
Second coaxial tree
required for 2x2 MIMO
Passive DAS (shared)
fulfilling the goal to have a fully coordinated radio-access network.
Today’s in-building system
Supporting users in indoor environments has been a challenge since the
start of mobile networking. For the last
two decades, this challenge has been
overcome by using a method referred
to as distributed antenna system (DAS).
The many flavors of DAS solutions are
all based on the principle of redistributing macro RBS radio signals across an
indoor antenna grid in the downlink,
and a corresponding collection of the
user traffic in the uplink. As illustrated
in Figure 1, this can be achieved by
using a passive coaxial RF distribution
network, or by using an active fibercoaxial hybrid network.
Distributed antenna solutions have
worked well for many years and are
still considered for multi-operator and
neutral host applications. However, the
technology becomes limited as requirements for higher capacity and capabilities increase and more advanced
services evolve. The DAS model originates from large-cell radio architecture,
and it is good for voice and basic data
coverage, but the radio bandwidth per
user it provides is too low to be a viable
solution as capacity needs rise.
E R I C S S O N R E V I E W • 2/2014
Indoor made simple
40
The uplink near-far problem – a UE connected to an outdoor macro degrades
SINR for the UE served by the indoor system
FIGURE 2
Power
Wanted
carrier
Blocking
carrier
ACLR
SINR
Indoor
antenna
RF frequency
Macro
connection
Blocking UE
Served UE
Indoor domain
The capacity challenge is of particular interest for mobile enterprise scenarios, as application usage shifts from
legacy laptop systems to smartphonebased consumption, which rapidly
increases indoor-radio capacity requirements. In many markets, the shift to
smartphone consumption has already
occurred for basic applications such
as e-mail, and is increasing rapidly as
major enterprise and consumer applications are adapted for smartphone usage.
Indoor radio challenges
Usage in indoor cellular environments
is shifting from traditional voice coverage to smartphone app coverage and
high performance mobile broadband.
For this transformation to succeed
and result in an immersive experience of nearly-instantly available data,
much higher capacity per unit area is
needed compared with existing solutions. However, with the high outdoorto-indoor penetration loss of modern
buildings, an improved indoor system is
E R I C S S O N R E V I E W • 2/2014
Outdoor domain
necessary. For other scenarios, advanced
outdoor macro cells with MIMO, carrier aggregation and beamforming are
suitable.
Pushing down the uplink receiver
noise level to a few decibels above the
thermal noise is a successful approach
to extend the reach of a macro radio,
but is useless for indoor radios. Instead,
dense antenna grids are necessary to
combat the uplink near-far effect –
spectral leakage from user equipment
(UE) near an indoor antenna, but connected to an outdoor macro cell and
transmitting on an adjacent carrier, can
substantially degrade uplink signal-tointerference-plus-noise ratio (SINR) of
the indoor node, possibly to the point
where service outage occurs.
This near-far effect is illustrated in
Figure 2 and cannot be mitigated by
filtering in the base station, as noise
from the blocking UE is inside the carrier bandwidth of the served UE. For
a 20MHz-wide LTE uplink carrier, the
maximum allowed spectral leakage
in the adjacent channel (ACLR) is 30dB
below the carrier power6. For example,
a UE transmitting at 1.7GHz to an outdoor macro cell using maximum transmit power, and assuming a distance
of 1m between the UE and the indoor
antenna, yields an SINR degradation
corresponding to an effective uplink
noise figure as high as 58dB.
The near-far effect does not affect
peak rates since it is not present all the
time. Once it is present, however, it puts
an upper limit on the coverage radius
per antenna due to the risk of service
outage. In the given example, service
outage could occur already with a coverage radius of 20m, depending on UE
capabilities and indoor propagation conditions. For such dense deployments,
fairly high levels of uplink noise can be
tolerated without performance degradation. Instead, focus should be placed on a
design that enables coordination as well
as fast and flexible deployment.
One particularly important requirement for the large building segment is
the need for tight coordination, to handle the dynamic traffic situations that
arise in complex radio environments
like modern atrium buildings with
open offices.
Attempts to apply the femtocell
model in such environments, which
lack natural cell borders, have proven
to be challenging. Instead of increasing capacity, reducing the cell size
often leads to reduced performance
and increased risk of dropped calls due
to inter-cell interference and frequent
handovers. At low loads, peak data rates
may become limited by control channel pollution, as each cell needs a dedicated and robust control channel. Thus,
deployment of femtocells creates a huge
challenge in terms of performance and
TCO. To maintain user satisfaction,
supporting interference management
and seamless mobility are crucial –
between the cells inside the building
and between outdoor and indoor cells.
This level of coordination is simply not
present in femtocell solutions.
The ability to add new features
through software upgrades, avoiding site visits as far as possible, is a key
success factor for indoor radio deployments. To ensure a consistent user
experience with full performance and
service functionality throughout the
41
network, having the same set of radio
features in the indoor segment as in the
outdoor macro RBS is desirable. This
also enables coordination between the
indoor and the outdoor environment
and simplifies network operations and
maintenance (O&M). Coherent QoS,
high-quality voice, and good mobility support, including for instance soft
handover for WCDMA, are examples of
features that will be important for user
satisfaction both indoors and outdoors.
As well as meeting requirements
for increased performance, enabling
large-scale rollouts requires a substantial reduction in complexity of installation. Reusing LAN cables is one key way
of achieving this. In addition, indoor
radio containing active equipment must
support remote powering, like PoE or
PoE+, as the need for local powering in
the event of a power outage could substantially increase deployment costs
and decrease availability.
Key design considerations
Given the challenges, new generations of indoor radio systems need to be
designed smartly, adopting best practices from existing indoor systems –
DAS, Wi-Fi, and Pico – and embracing
new features.
Feature parity
To achieve the desired performance
gain from cell size reduction, combined
cell technology with spatial division
multiple access (SDMA), coordinated
scheduling and other advanced coordination features are needed. Such features are already available in macro
environments, and sharing the same
software base for indoor and outdoor
radios greatly simplifies the implementation of feature parity.
A convenient approach is to use
the same family of DU, which is also
referred to as the baseband unit. DAS is
based on such a design using the same
hardware and software to drive all
antennas in both the indoor and the
outdoor macro network.
Fronthauling with LAN cables
To facilitate deployment of smaller cells
with full coordination and scalability
for high capacity, a fronthaul architecture with a star topology is desired. This
approach enables each radio head to be
FIGURE 3
Radio Dot System – indoor made simple
Radio Dots
Radio Dots
IRU
DU
Structured LAN cabling
Remote powering
Radio base station
with DU and IRU
Power and
backup
fronthauled individually, resulting in
an indoor radio solution that is capable of supporting high capacity while
retaining maximum flexibility. The use
of LAN cables means that several indoor
systems can be deployed within the
same budget and time limitations as is
required for a typical DAS – as the traditional method of fronthauling requires
a fiber deployment. The design challenge in this scenario is how to fronthaul effectively through LAN cables.
Form factors
One concept related to the Internet of
Things is that of miniaturized design
or integration of communication into
everything in a natural way. For our
indoor system design, an ultra-compact
form factor for the radio heads was one
of the most important design considerations. To achieve compactness, low
power design is essential so that the
heat can be dissipated without affecting equipment reliability. The target
is a compact radio head that is smaller
than current DAS antennas, with a
minimalist design suiting any indoor
environment.
Support for high bandwidth
To meet ever-increasing capacity
demands, high bandwidth is essential
for high-performance radio systems,
Radio Dots with
MIMO and diversity
which can be achieved through carrier aggregation in wide FDD and TDD
bands. Additional benefits will come
from the adoption of 4x4 MIMO, which
should occur in the near future, doubling the total bandwidth capacity.
A novel indoor radio solution
As a new generation of indoor radio
systems, the RDS has been developed
with these key design considerations
in mind, based on technology already
developed for RBSs, but with a focus
on reducing architecture complexity, enhancing system scalability and
improving radio system performance.
The design utilizes LAN cabling infrastructure to connect the active antenna
elements – which are called Radio Dots.
The system specifically targets use
cases that are demanding in terms of
performance and dynamic capacity allocation – scenarios that typically require
multi-antenna grids, such as mediumto large-size office buildings. For areas
with less demanding capacity requirements, such as parking garages or outdoor areas in a campus environment,
the RDS can often be complemented
with remote radio unit (RRU)-based
micro DAS.
As Figure 3 shows, the RDS has three
key components: the Radio Dot, the
indoor radio unit (IRU) and the DU.
E R I C S S O N R E V I E W • 2/2014
Indoor made simple
42
Digital unit
The DU provides pooled baseband processing for the system. To manage the
connected radios, the DU uses the CPRI
standard for the DU-IRU interface to
transfer synchronization, radio signals
and O&M signals. When collocated with
the IRU, an electrical CPRI interface is
used, and for remote connection with
the IRU, a CPRI fiber interface is used.
Indoor radio unit
A newly designed RU that incorporates existing macro software features,
extending them with new indoor features. The IRU connects to each Radio
Dot using a proprietary IRU-Radio Dot
interface over a LAN cable (detailed further on).
Radio Dot
The Radio Dot has two integrated antennas in a 10cm form factor and weighs
under 300g. Each Radio Dot is connected to the IRU through a dedicated
LAN cable and remotely powered by
PoE. As in Ethernet networks, the system employs a star topology, as opposed
to the tree topology used in DAS. The
ultra-compact design and use of LAN
cabling simplify installation.
The design of the system is essentially
a centralized baseband architecture
that enables baseband resource pooling
and full coordination. Initial baseband
capacity can be selected to meet nearterm demand, and more capacity can
be gradually added at the DU and IRU
as traffic demand increases –without
rates, as control channel pollution can
be avoided. The combined cell approach
can be taken one step further by introducing SDMA, which allows resources
to be dynamically reused within the
cell. This enables instantaneous scaleup to full capacity, while minimizing the
mobility overhead.
any need to modify the cable infrastructure and the installed Radio Dots. To
illustrate this point, a single cell per
IRU can be upgraded to a multiple cell
per IRU simply by exchanging the IRU
– leaving the Radio Dots untouched. In
addition, a 4x4 MIMO can be supported
without changing the installed cable
infrastructure.
The system is manageable all the way
up to the antenna element. The radio
properties of each individual Radio Dot
can be tuned in terms of coverage and
performance. The approach used in this
solution reduces the need for careful,
tedious and costly site investigations
and network planning for each building, by applying rule-of-thumb-based
network planning to define the deployment requirements such as inter-radio
distance per building type (based on
statistical simulation results for typical
floor plans). This approach simplifies
the planning process and is sufficient
to guarantee high performance for most
buildings. If needed, additional radios
can be installed at a later stage, and this
can be completed quickly due to the simple deployment and LAN-like star topology of the system architecture.
The ability to apply combined-cell
technology to the maximum results in
fully coordinated cells, which in turn
further optimizes capacity, mobility
and robustness of the indoor radio network. Combined cell minimizes the
number of handovers by allowing multiple radios to share the same physical cell identity. It also increases peak
FIGURE 4 Main-remote RBS block diagram
Remote radio unit
Digital unit
RF front-end
RF
IF
TX
DAC
Radio
Processing
RF
Duplexer
E R I C S S O N R E V I E W • 2/2014
RX
IF
ADC
CPRI
BB
processing
The cable interface
Figure 4 illustrates the basic block diagram of a conventional main-remote
RBS solution. Adaptation of this solution for indoor environments was a key
design goal of the RDS, and our idea was
to utilize cost-effective LAN cables to
enable a totally fronthaul-based architecture. The challenge then, was to
identify where the LAN cable interface
should be introduced.
Using existing Ethernet technologies,
such as 1000BASE-T or 10GBASE-T, to
transport CPRI over LAN cables is one
way to answer this question. However,
Ethernet PHYs with IEEE1588v2 support
were not originally designed to support
CPRI with stringent requirements for
bit-error-rate, latency and delay variation. This results in a compromise
between bit rate, reach and latency. To
address this, significant CPRI frame
compression is needed, which increases
complexity but also power consumption. Currently, the combined processing and compression for Ethernet PHYs
and CPRI result in relatively high power
consumption, and so due to the heat dissipation, this approach is not a good fit
for a slim design.
So, what other options are available?
LAN cables are capable of transporting
very high bandwidths; for example, the
effective bandwidth for 100m of Cat 6a
per used twisted pair is between DC and
400MHz. This high bandwidth is feasible, as it operates in the lowest part of the
spectrum and has both a low noise floor
and rather low cable loss. As four twisted
pairs are available for each LAN cable,
the question now becomes: is it possible
to efficiently exploit this bandwidth and
the four pairs for fronthauling?
If we take another look at the RU
design in Figure 4, an interface with a
low intermediate frequency (IF) exists
between the ADC/DAC blocks and the
down-/upconverters. So, is it feasible to
transmit the IF signals directly over the
LAN cables? The answer is yes. As shown
43
FIGURE 5
Radio Dot System block diagram
Radio Dot
Indoor radio unit
Digital unit
RF front-end
RF
TX
DAC
Cable
I/F
RF
Radio Dot interface
Cable
I/F
RX
Radio
Processing
CPRI
BB
processing
ADC
Duplexer
in Figure 5, such an IF-based design,
in effect, extends the RF front-end
over a LAN cable using an IF interface
– which transports the radio signals at
low frequencies with graceful capacity
degradation in the event of unexpected
cable faults.
The elegance and simplicity of this
design requires minimal hardware/software changes regarding radio front end
and processing, and enables the overall
ultra-compact design (see Figure 3). The
IF-based design requires a lot less power
compared with possible Ethernet-based
methods, as the IF cable interface can be
designed in a more power-efficient way.
In addition to the radio signals, the same
twisted pair can carry synchronization
signals and control signaling and power.
This design also supports advanced features for cable equalization, AGC, cable
testing and troubleshooting.
The IF-based design provides high
radio performance and support beyond
the standard LAN-cabling reach of
100m. Given the low noise floor and
rather low cable attenuation at selected
IF frequencies, 3GPP uplink and downlink requirements can be fulfilled, and
4x4 MIMO can be supported by utilizing all four pairs of the cable. Previous
research has shown that the use of four
antennas for indoor environments
has great potential to further increase
capacity7.
Copper is the medium of choice for
indoor broadband infrastructure and
will remain so for many years. LAN
cable technology has evolved significantly over the past four decades – from
Cat 3 to Cat 7 and the upcoming Cat 8
– driven by improvements in Ethernet
speeds from 10Mbps to 10Gbps and set
to reach 40Gbps in the near future. This
has led to substantial improvements in
cable technology, which offer higher
FIGURE 6
bandwidth and lower noise. The RDS
will continue to build on this evolution,
both for performance upgrades and cost
erosion due to economies of scale.
Lab test results
The performance of the system has
been verified in lab tests. Figure 6
System performance
DL throughput (Mbps)
350
Test samples
Fitted
300
250
200
150
100
–5
0
5
10
15
20
25
SINR on SCC (dB)
E R I C S S O N R E V I E W • 2/2014
Indoor made simple
44
FIGURE 7
Illustration of flexible capacity
Evenly configured
More capacity
in hotspots
Hotspot
shows the result of a DL test with
carrier aggregation of two 20MHz LTE
carriers in the 2.1GHz band and using
2x2 MIMO. The IRU and Radio Dot were
connected using 190m of Cat 6a cable.
The SINR on the primary component
carrier (PCC) was fixed at 27dB to show
the full peak rate of carrier aggregation,
while the SINR on the secondary component carrier (SCC) was varied between
-5 and 25dB. During the test, the DL
throughput on the PCC maintained the
expected peak rate of 150Mbps throughout. Figure 6 shows the aggregated DL
throughput versus the SCC SINR, where
the throughput increase above 150Mbps
is due to the SCC. The aggregated peak
rate of 300Mbps was achieved at about
23dB SINR.
Evolution to flexible capacity
Indoor traffic demand tends to vary
over time and space, particularly in
enterprise and public environments.
For example, traffic demand regularly
increases over the course of a day in
areas where many people gather, such
as in conference rooms, cafeterias, and
lobbies. This high traffic demand disappears once people leave. Evenly distributing high capacity in a building for
its peak use is not the best approach, as
this tends to result in overprovisioning
capacity.
E R I C S S O N R E V I E W • 2/2014
Coverage only for
energy efficiency
Hotspot
As the RDS uses centralized baseband
architecture, it can provide capacity in
a more flexible way – by shifting available capacity from one place to another
on demand. This can be implemented
through dynamic cell reconfiguration (such as, traditional cell splitting
and combining) or by using combined
cell SDMA technology. For LTE Rel10/11 UEs, combined cell SDMA is the
desired approach for dynamic SDMA
operations in one cell involving all the
radios. This approach enables efficient
use of the available baseband capacity,
optimizing both network capacity and
mobility, resulting in an improved user
experience. Overlapping radios can be
turned off (dynamically) to save energy.
Figure  7 shows three typical scenarios assuming three-cell baseband capability. Here, for illustration purposes
only, a dynamic cell reconfiguration
approach is used.
In the first scenario, three cells are
distributed evenly to cover the indoor
area, and each cell contains five radios.
The second scenario covers the same
space but includes two traffic hotspots.
Here, the top cell is split into two smaller
cells to provide higher capacity to the
hotspots, while the rest of the area is
covered by a single larger cell using the
remaining baseband resources. In the
third scenario, traffic demand is very
low – a common situation late at night
and early in the morning. To provide
capacity for this low traffic scenario, the
orignal three cells are combined into one
large cell with only the selected radios
active. All other radios (including the
baseband resources involved) are inactive to save energy.
Summary
In this article, we have highlighted the
challenges related to radio capacity and
performance inside buildings, summarizing the main requirements to be
successful in overcoming them. With
a limited technology toolbox available
to operators today, scalable growth for
the platforms of the Networked Society
is restricted, and so innovative design
principles for smart and flexible small
cell radio technology are needed.
Our aim was to provide operators
with the best combination of two worlds:
superior radio technologies and their
continual evolution from the mobile
industry, together with the well-established LAN building practices of the IT
community. This was our inspiration
for the design of the Radio Dot System
– a novel indoor small cell solution.
45
Chenguang Lu
Kim Laraqui
Olle V. Tidblad
is a senior researcher in small
cell transport within Ericsson
Research and is part of the
research team developing the
RDS concept. He joined Ericsson in 2008
and holds a Ph.D. in wireless
communications from Aalborg University,
Aalborg, Denmark. He has actively
contributed to DSL technologies like
Vectorized VDSL2 and G.fast. Since 2010,
he has mainly focused on research in small
cell backhauling and fronthauling.
is a principal researcher and
technical driver of research on
transport solutions for
heterogeneous networks. He is
also part of the research team developing
the RDS concept. He joined Ericsson in
2008 as a regional senior customer solution
manager. Prior to this, he was a senior
consultant on network solutions, design,
deployment and operations for mobile and
fixed operators worldwide. He holds an
M.Sc. in computer science and engineering
from KTH Royal Institute of Technology,
Stockholm, Sweden.
joined Ericsson in 1996 as
field implementation
supervisor of fiber access and
transmission solutions. From
1997 to 2000, he worked with enterprise
communication and IP solutions at the
international carrier Global One. Since
returning to Ericsson, he has held product
management positions within fixed and
radio access infrastructure including
WCDMA, LTE and small cells. He has worked
with Ericsson research and radio product
development to bring the RDS to market. He
holds a B.Sc. in electrical engineering, and
applied telecommunication and switching
from KTH Royal Institute of Technology,
Stockholm, Sweden.
Henrik Almeida
joined Ericsson in 1990 and is
currently head of Small-Cell
Transport at Ericsson
Research, where the RDS
concept was developed. He has a long
history of working with fixed-line access
technologies at Ericsson and is now
focusing on small cell transport solutions for
the Networked Society. He holds a Ph.D.
H.C. from Lund University, Sweden, for his
work in the area of broadband technologies.
Elmar Trojer
joined Ericsson in 2005 and is
currently working as a master
researcher in the Small-Cell
Transport group at Ericsson
Research. He is part of the research team
developing the RDS concept. He has worked
on both fixed and mobile broadband
technologies such as VDSL2 and G.fast,
dynamic line management solutions for
IPTV, and 3G/4G access in the context of
mobile backhaul/fronthaul over copper and
fiber media. He led the design and product
systemization of the RDS with a strong
focus on the physical layer radio
transmission. He holds a Ph.D. in electrical
engineering from the Vienna University of
Technology, and an MBA from the University
of Vienna.
Miguel Berg
joined Ericsson Research in
2007 and is currently a master
researcher in the Small-Cell
Transport group. He is part of
the research team developing the RDS
concept. From 2007 to 2011, he was active
in research regarding copper cable
modelling and line-testing algorithms for
xDSL and G.fast. He holds a Ph.D. in
wireless communication systems from
KTH Royal Institute of Technology in
Stockholm, Sweden. Between 2002 and
2003, he worked at Radio Components,
where he was involved in the design of base
station antennas and tower-mounted
amplifiers.
Per-Erik Eriksson
joined Ericsson in 1989. He is
currently a senior researcher at
the Small-Cell Transport group,
Ericsson Research. He is part of
the research team developing the RDS
concept. He has previously worked with
ADSL, Vectorized VDSL2 and G.fast and has
also been involved in the standardization of
those technologies. He holds an M.Sc. in
electronic engineering from KTH Royal
Institute of Technology, Stockholm,
Sweden.
References
1. Ericsson Mobility Report, June 2014, available at:
http://www.ericsson.com/res/docs/2014/ericsson-mobility-report-june-2014.pdf
2. United Nations, 2014 Revision, World Urbanization Prospects [highlights], available at:
http://esa.un.org/unpd/wup/Highlights/WUP2014-Highlights.pdf
3. Cisco, 2014, Visual Networking Index: Global Mobile Data Traffic Forecast Update, 2013–2018,
available at: http://www.cisco.com/c/en/us/solutions/collateral/service-provider/visualnetworking-index-vni/white_paper_c11-520862.html
4. CPRI Specification V6.0, 2013-08-30, available at:
http://www.cpri.info/downloads/CPRI_v_6_0_2013-08-30.pdf
5. Ericsson Review, June 2014, 5G radio access, available at:
http://www.ericsson.com/news/140618-5g-radio-access_244099437_c
6. 3GPP, Technical Specification 36.101, LTE; E-UTRA; UE Radio Transmission and Reception,
version 11.8.0, available at: http://www.3gpp.org/dynareport/36101.htm
7. IEEE, 2013, LTE-A Field Measurements: 8x8 MIMO and Carrier Aggregation, Vehicular
Technology Conference (VTC Spring), abstract available at:
http://dx.doi.org/10.1109/VTCSpring.2013.6692627
E R I C S S O N R E V I E W • 2/2014
Programmable networks
46
Architecture evolution for
automation and network
programmability
The target architecture of future telecom networks will be designed using sets of aggregated
capabilities. Each domain will have its own set of resources that are abstracted and exposed to other
domains, supporting multi-tenancy and tenant isolation. The result is a fully programmable network,
that has the ability to evolve and adapt to the emerging requirements of the Networked Society.
GÖR A N RU N E , E R I K W E S T E R BE RG, T OR BJÖR N C AGE N I U S ,
IGNAC IO M A S , BA L Á Z S VA RGA , H E N R I K BA S I L I E R , A N D L A R S A NGE L I N
Enabled by emerging
technologies like virtualization,
software-defined networking
(SDN) and cloud capabilities,
the architecture of telecom
networks is undergoing a
massive transformation. This is
being driven by several factors,
including the need for less
complex and lower-cost network
operations, shorter time to
customer (TTC) and time to
market (TTM) for new services,
and new business opportunities
built on the anything as a service
(XaaS) model.
The principles of the target architecture
are based on separation of concerns,
multi-tenancy and network programmability. As networks progress toward
the target architecture supporting
as-a-service models with rapid scalability capabilities and greater levels of automation, the need to focus on the basic
principles will become more significant.
Full programmability of a network
and its services needs to take all the
building blocks of a network into consideration: how each piece will evolve;
how they will interface; and how they
support the structure and business processes of an operator.
SDN technologies, for example, are
key enabling tools for network programmability, but to provide value they must
be integrated with the end-to-end process view of the operator. Cloud orchestration technologies are also important
enablers, but without proper interfaces
to business management functions in
place, the result would be a technically
functional but commercially dysfunctional system. Well-defined technical
BOX A Terms and abbreviations
AAA
authentication, authorization and accounting
API
application programming interface
APN
Access Point Name
BSS
business support systems
COMPA
control, orchestration, management, policy and analytics
DC
data center
EPC
Evolved Packet Core
IGP
Internet Gateway Protocol
IPS
infrastructure and platform services
MPLS
multi-protocol label switching
MTC
machine-type communication
MVNO
mobile virtual network operator
E R I C S S O N R E V I E W • 2/2014
NFV
OSS
opex
OVF
PaaS
POD
R&S
SDN
SLA
TTM
TTC
VM
vDC
VIM
network functions virtualization
operations support systems
operational expenditure
Open Virtualization Format
platform as a service
performance-optimized data centers
routing and switching
software-defined networking
Service Level Agreement
time to market
time to customer
virtual machine
virtual data center
Virtualized Infrastructure Manager
interfaces and abstractions are critical
to facilitate a split of responsibilities,
support trust relationships and enable
opex efficiency.
This article aims to describe the big
picture of the target ecosystem, presenting an architecture description that
focuses on the inter-domain interfaces,
separation of concerns as well as network programmability.
The ecosystem
The target network architecture will
be built using a set of critical technical interfaces that support business
relations – which we call inter-domain
interfaces. These interfaces mark the
boundaries between the different layers
or domains of a network; they support
the separation of concerns, interoperability, and enable Service Level
Agreements (SLAs). Administrative
domains, as defined by NFV1, are suitable for being managed as one entity
from a competence and administrative
responsibility point of view. As Figure 1
illustrates, there are four typical administrative domains:
transport;
infrastructure and platform services ;
access and network functions; and
business and cross-domain operations.
The target architecture – and in particular the inter-domain interfaces – serve
as enablers for a multitude of domain
combinations. Many other domain
structures are possible, depending on
the strategy and operational structure
of the operator.
Administrative domains are quite
physical in nature. Traditionally, they
47
tend to consist of physical nodes with
pre-integrated hardware and software
functions. This, however, is changing.
Together, NFV and the separation of
software and hardware have brought
about a new administrative domain:
the infrastructure and platform services (IPS) domain. Some administrative
domains – notably transport, access network and the new IPS domain – maintain responsibility for hardware and
platforms, while most other network
function domains – such as the Evolved
Packet Core (EPC) – manage only software functions.
Even though current network architecture already includes several interdomain interfaces, the evolution to the
target architecture aims to improve
multi-tenancy capabilities, as well as
intra-domain and inter-domain programmability. This evolution will happen gradually and to varying degrees
for each domain depending on need –
in terms of value – as well as additional
considerations like legacy equipment
and operational processes.
Key principles of the target
architecture
Developing network architecture so
that it is both highly automated and
programmable requires functionality
to be coordinated across administrative
domains. This can be achieved through
a set of tools to operate each administrative domain, which have operational responsibility for the resources
within the domain, as well as the ability to expose services based on these
resources. In this article we refer to the
combination of these operational tools
as COMPA: control, orchestration, management, policies and analytics. Each
term has a wider meaning than its legacy definition; all are tightly interlinked
within each administrative domain, as
well as having inter-domain relations.
The COMPA functional groupings are
illustrated in the target architecture
shown in Figure 2.
The main principles of the target
architecture are:
separation of concerns;
abstraction and exposure of capabilities;
multi-tenancy;
intra-domain programmability; and
inter-domain programmability.
FIGURE 1 Target architecture with example administrative domains
Access
Core
Services
Business
and cross-domain
operations
Access and network functions
Control
Orchestration
Management
Policies
Analytics
Infrastructure and platform
Transport
Control, orchestration and management
Management and control functions
within each domain will do much the
same job as they do today, but with a
higher degree of automation and realtime capabilities. Orchestration enables
automation across different types of
resources and uses defined workflows
to provide the desired network behavior – all aligned with and enabled by a
policy framework that is supported by
analytics insights. Creating infrastructure services is one example of where
orchestration is heavily used in the IPS
domain, in which processing, storage
and networking resources are assigned
in a coordinated manner.
Services from other domains can also
be viewed as resources orchestrated in a
synchronized manner with a domain’s
own resources to provide services in a
hierarchical way. A strict framework
with a common information model is
required to maintain consistency across
domains – illustrated by the verticalarrow flow in Figure 2.
FIGURE 2 Grouping of COMPA functions in the target architecture
Access and network
functions
Infrastructure
and platform
Transport
Control
Orchestration
Management
Policies
Analytics
Control
Orchestration
Management
Policies
Analytics
Business and
cross-domain
operations
Control
Orchestration
Management
Policies
Analytics
Control
Orchestration
Management
Policies
Analytics
E R I C S S O N R E V I E W • 2/2014
Programmable networks
48
FIGURE 3 Policy framework
Operator level
Strategic, tactical and commercial policies:
Business and cross-domain operations
Policy administrative domains
System level policies per
administrative domain
Network functions (NF groups)
Detailed policies:
Network functions level
To offer services that draw
resources from more than one domain,
a cross-domain OSS/BSS function is
needed. This second main flow of
orchestration relates to external business offerings and how to leverage
services from multiple domains. For
example, an enterprise customer may
require a service that combines an infrastructure service from the IPS domain
with a business VPN from the transport
domain – this is shown conceptually by
the horizontal arrows in Figure 2.
To support service exposure, each
domain needs appropriate logging tools.
For example, an IPS domain will need
to create and maintain data records
related to usage for the infrastructure services it provides – regardless of
whether it delivers these services to an
external tenant or to an internal tenant (to other domains within the same
operator). Many of these functions will
be automated and simplified in their
interfaces among staff, OSS/BSS, and
resource control functions.
The policy framework
Policies are sets of rules that govern
the behavior of a system. A policy comprises conditions and actions; events
trigger conditions to be evaluated and
actions are taken if the conditions are
met. Policies are used to define a framework and set the bounds for the controlorchestration-management functions,
derived from the overall business goals
of the operator.
E R I C S S O N R E V I E W • 2/2014
Some policies, like those that control how specific resources are used,
are strictly defined and applied within
an administrative domain. Other policies apply to the inter-domain interfaces, and define for example how one
domain can use services from another.
Such policies can be partly defined by
the administrative domain delivering
the service, but may also be defined by
the administrative domain for business
and cross-domain operations. Figure 3
shows how policies originate from the
overall business objectives of the operator and how they relate to different levels within the operator structure.
The relationship between business
and network operations policies is
defined by a set of meaningful operational KPIs. For example, a business policy governing the parameters of a gold
subscriber service can be interpreted
into specific settings for, say, QoS in the
network. By factoring in the insights
supplied by analytics, these operational
KPIs enable a greater degree of network
automation, and allow policies to govern operational decisions.
Network analytics
Analytics is therefore a key tool for
increasing automation of operations.
To provide insights, predictions, as
well as supporting automation in other
ways, analytics can be applied within an
administrative domain or work in conjunction with the other COMPA functions – both in offline processing of data
and for real-time stream processing.
Domain competence is usually needed
to understand prediction, but insights
exposed from other domains or external sources could also be used as input.
Exposing analytics insights on a
domain basis, and then aggregating multiple domains through a cross-domain
analytics application, enables the entire
network state to be analyzed; which in
turn supports the definition of networkwide KPIs.
A policy engine can use network analytics to check performance-related
KPIs, triggering network state updates
when needed. Such requests could
then be applied to the relevant network
domains by the control-orchestrationmanagement functions – possibly with
some form of manual intervention.
A closed feedback loop from the control-orchestration-management functionality back to the policy engine
would enable policies to learn and adapt
automatically as the network environment changes.
Applying the concepts
Transport
In telco networks, the transport domain
delivers connectivity services between
remote sites and equipment, maintaining topology awareness and services for
multiple customers – multi-tenancy.
In reality, a transport network consists of a set of interworking administrative domains defined by technology,
geography and ownership. The main
technologies powering the delivery of
connectivity services will be based on
IP/MPLS, Ethernet and optical transport;
in the access domain, microwave transport may also play a significant role, and
IPv6 will be the dominant protocol (as
IPv4 becomes more associated with legacy infrastructure). Transport network
topology will become flatter with fewer
packet hops, as the use of converged
IP and optical transport technologies
becomes more widespread2.
Traditional connectivity services like
residential broadband, mobile backhaul, and enterprise VPNs will coexist
with newer services that will provide
connectivity for cloud solutions, such as
DC-to-DC or user-to-DC. These new generation services and the increased number of connections will drive the need
49
for more flexible and dynamic ways to
operate the transport domain.
A number of key components are
needed to support evolved architectural principles and facilitate both
intra-domain and inter-domain programmability. These components
include SDN and network virtualization
technologies3, which allow connectivity
services to be deployed and controlled in
a flexible way.
Programmability in the transport
domain will ensure a suitable level of
resource abstraction, exposure and
control so that other administrative
domains can request transport services according to established SLAs.
Programmability can be achieved by
using northbound SDN-based interfaces, for example, and can be further
increased by leveraging the benefits of
data/control plane separation.
As shown in Figure 4, several scenarios regarding what parts of a transport
node can be SDN controlled. These scenarios lead to multiple possible paths and
intermediate steps to transform a traditional transport network into a network
that is fully SDN-controlled – in which
only a limited set of functions are local
to the transport node. Using SDN controllers will not only result in the introduction of new functions and services
into transport nodes, but existing control functionalities will be moved to the
SDN controller – replacing current localnode implementations.
Migrating an existing transport network to an SDN-based architecture
requires hybrid operational modes
that apply SDN-based control capabilities onto the existing (protocol-driven
local node) transport infrastructure. The
capabilities that are included depend on
the level of centralization versus distribution of functions that the operator
chooses for its transport domain.
The resulting transport domain – in
the context of packet-optical integration – combines increased programmability (enabled by SDN technologies) with
the simpler, more cost-efficient IP and
optical components, and is detailed in
a previous Ericsson Review article2. The
evolved transport domain enables faster
service deployment and reduces operational complexity.
Infrastructure and platform services
As networks evolve, telecom solutions
and systems will increasingly be built
using on-demand elastic infrastructure
and platform services rather than dedicated and managed infrastructure and
software. To leverage the benefits of this
model, a split in responsibility between
the provider of such services and the
users (tenants) is necessary. The provider role is taken by what we refer to in
this article as the IPS domain, which is
a new domain type that provides infrastructure and platform services using
owned or leased resources.
One of the key services offered by the
IPS domain is a structured collection of
virtual computational processing, storage and networking resources, within
what is referred to as virtual data center
(vDC). The vDC interface separates logical telecom nodes from the actual physical infrastructure, using concepts like
virtual machines, virtual network overlays, baremetal, and storage services.
Networking capabilities exposed to
tenants will be rich enough to support
a wide set of telco functions, including
L2 and L3 VPN interworking and SDNcontrolled service chaining4. The IPS
domain can also take the administrative responsibility for common network
functions (such as DNS, firewalling,
DHCP, and load balancing) and offer
these as services, orderable as products
deployable in a vDC.
In addition, the IPS domain can also
supply services to applications, providing an execution framework (PaaS) and
network APIs that expose underlying
network capabilities. For example, common network functions can be exposed
and made programmable by applications. Inter-domain programmability
and abstraction increases application
development productivity and reduces
lead times. In addition, the IPS domain
will support migration by providing
interconnectivity with non-virtualized networks as well as mixed
FIGURE 4 Scenarios for control plane and data plane separation for packet,
and IP/optical transport networks
SDN
controller
SDN
controller
Service
Transport
Service
function
Service
SDN
controller
Optical
SDN
controller
Optical
Service
Transport
Service
Service
function
Service
function
Service
function
Transport
function
Transport
function
Transport
function
Transport
function
Port
Port
Optical
Optical
Hybrid SDN
legacy mode
(packet)
Full SDN mode
(packet)
Hybrid SDN
legacy mode
(IP+optical)
Full SDN mode
(IP+optical)
IGP
IGP
E R I C S S O N R E V I E W • 2/2014
Programmable networks
50
FIGURE 5 Infrastructure and platform services domain
Tenant domain
Overall IPS
functions
Infrastructure resource
zone or provider
COMPA
VIM
(v)
switch
Virtualization
POD
App
framework
HW/OS
External
infrastructure
service
provider
POD
DC
interconnect
Data site center 1
deployments of non-virtualized, virtualized and PaaS-based applications.
All the capabilities of the vDC and
application services are orderable by
tenants through policy-controlled
inter-domain interfaces, and all of the
capabilities can be requested, monitored and maintained/scaled through
these interfaces. The interfaces will rely
heavily on modeling of the (sometimes
complex) sets of capabilities, using OVF
descriptors, for example, and forwarding descriptors for service chaining.
Within the IPS domain, overall functions in the COMPA category will act
across a wide set of resources in the
underlying infrastructure.
Using orchestration technologies,
for example, suitable abstractions can
be provided to tenants using a heterogeneous set of resources – which
allows tenants to manage and program resources without requiring any
lower level implementation details.
Policies and analytics may then be
used to ensure that resources are used
E R I C S S O N R E V I E W • 2/2014
Platform
resources
POD
Transport
POD
DC
interconnect
Data site center 2
efficiently, while respecting SLAs and
business requirements.
The physical resources that expose
virtual resources to tenants may be
organized into infrastructure resource
zones, each with their own functions
(VIM in ETSI NFV terminology) acting
within the zone – such as OpenStack
and SDN controllers. Some or all such
zones may be external to the IPS
domain. Another option is to use similar services from another IPS domain
or service provider, where orchestration
capabilities deliver a consolidated service. The transport domain may be used
for inter-connectivity of infrastructure
resource zones at different data center sites or to connect infrastructure
resource zones to external networks.
In both cases, the IPS domain interacts
with the transport domain, based on
frame agreements, to request or dynamically adapt WAN connections.
As shown in Figure 5, the IPS domain
relies on several arbitrarily distributed
DC sites, which contain a number of
PODs – blocks of computational, storage
and networking resources. Typically, a
POD corresponds to an infrastructure
resource zone. To deliver consolidated
and distributed vDCs, the overall orchestrator can request resources across the
PODs through their VIM functions.
The IPS domain offers abstracted
services (the vDCs and application services), multi-tenancy with isolation of
resources, security and SLAs associated
with these services. It allows for intradomain programmability and automation via the VIM (OpenStack), SDN
for the connectivity resources and the
COMPA functions for resource and service orchestration across infrastructure
resource zones and to external providers. It also offers inter-domain programmability where tenants have access to
interfaces for controlling – within
frame agreements – their instances of
the vDC and application services, supporting for example scaling, tenant
SDN control or access to telco network
capabilities. The interface between the
IPS domain and its tenants needs to be
open and, where applicable, standardized to support a full business ecosystem
between IPS-domain service providers and its tenants, with a minimum
amount of system integration between
the two. Indeed, this appears to be one of
the main tasks of the NFV forum.
Network functions
Most network functions of the logical telecom architecture shown in
Figure 6 benefit from using services
from the IPS domain. The separation of
network functions from platforms can
result in significant operational gain –
primarily through automated routines
for backup and restore, capacity planning, hardware handling and a general
reduction in the number of platforms
to be managed. This has a direct impact
on TTM for new services, which can be
reduced from up to a year down to a few
months as the introduction process no
longer depends on platform introduction. Auto scaling of the infrastructure
and platform services and programmability of the network functions removes
much of the manual work associated
with fulfillment, which greatly reduces
the TTC.
The original design of mobile network
architecture in 3GPP supports a certain
51
level of programmability, abstraction
and multi-tenancy. Standardized interfaces between the RAN, EPC and IMS
domains support automation in bearer
service handling and a set of MVNO solutions at various levels. The Rx interface
enables rudimentary inter-domain programming to the PCRF from outside the
EPC domain, while the APN structure
provides a foundation for multi-tenancy.
However this is not sufficient, network
functions architecture is evolving to
increase support for COMPA functions.
Introducing the infrastructure and platform services are a significant step in
this direction, but additional architectural changes and interface improvements are also part of the wider picture.
Separating network functions from
the platforms allows the capacity of a
given network function system – such
as an EPC system – to scale up or down
by simply adjusting the capacity of the
vDC to achieve the wanted capacity of
the EPC system. The multi-tenancy of
the vDC service also means that multiple EPC systems can be instantiated in
parallel in separate vDCs.
Figure 7 illustrates how deploying a
multitude of EPCs in different vDCs provides full isolation of the EPC instances,
inherited from the tenant isolation
built into the vDC service from the IPS
domain. Isolation makes both service
exposure and inter-domain programmability to EPC instances safer – opening up programmability to one instance
does not impact others, and exposure
of data from the EPC system to a customer or partner is limited to that of
the associated EPC system instance.
Implementing isolation in this way minimizes risk and reduces the cost for troubleshooting faulty services.
For operations in multiple markets,
one EPC system can be instantiated per
market, with a central responsibility for
the EPC domain, but with selected programmability suitable for the demands
of the given market. This is a cost-efficient approach with consolidated competence and responsibility, while still
allowing different operational entities
to control selected features of the EPC
system – such as rules for charging or
subscription.
Instantiating a VoLTE system5,
for example, can enable an operator
to offer communication services to
enterprises, emergency services or any
other industry with full isolation and
varying degrees of programmability. To
support this use case, network architecture needs to evolve to the target
architecture. In particular, additional
inter-domain interfaces (to enable programmability and automated orchestration) are needed to instantiate the
relevant subsystems and combine them
into service solutions.
The evolution of the network functions integrates well with 5G radio evolution6. Next generation networks will
support legacy services as well as new
services like enhanced mobile broadband, massive machine-type communication (MTC), as well as mission-critical
MTC. Future networks will need to
support a vast number and a much more
diverse set of use cases. Consequently,
service creation that is platform-independent and flexible, based on programmability and automation is key. A
massive range of industries will depend
on 5G networks – all with different
requirements for characteristics, security, analytics and cost. Meeting all of
these needs is a strong driver for multitenancy, isolation, and instantiation of
services and resources.
Extending instantiation capabilities
to work across multiple domains may
enable novel business offerings to be
created. If, for example, an instance
of an EPC system is integrated with a
VoLTE system instance, the two are then
connected to an IP VPN, and finally
FIGURE 6 Logical telecom architecture
OSS/BSS
User data management
AAA
HSS/
HLR
Domain
mgmt.
OSS
BSS
UDR
Service
enablement
Expose
Packet core
Communication services
PCRF
Mobile
CS
S/PDN
GW
MME/
SGSN
eMBMS
GW
ePDG
GW
TDF
SCCF
IMS
core
IMS
IMS
telephony
message
Media services
Media
delivery
Mobile
broadcast
Other services
Server
Server
E R I C S S O N R E V I E W • 2/2014
Programmable networks
52
all three are associated with an isolated and SLA controlled radio-access
service; the result is an isolated, and
SLA-controlled logical instance of the
complete network. Such logical network
instances can be offered to an industry,
to an MVNO or an enterprise. As each
network instance is isolated, it is safe
to open up interfaces to each instance
to enable each customer or partner to
program selected properties of the logical network instance, and to do this in
real time.
To reach the point where a network
can be offered as a programmable service requires a cost-efficient way to connect services – and eventually resources
– from the various domains into logical network instances. As described at
the beginning of this article, to connect services in such a cost-efficient
way requires inter-domain programmability and more generally a networkwide architecture for cross-domain
orchestration and management, while
maintaining per-domain responsibility
and accountability.
many network functions will be managed in similar way as any other virtualized software: following virtualization
management principles in line with
ETSI NFV specifications. Initially virtualized network functions will be operated in parallel with legacy nodes, and
DC operations as well as maintenance
will be automated to a much larger
degree than it is today.
In the longer term, the architecture
should be able to provide the desired
level of automation and network programmability. Full programmability of
the network and its services requires
the inter-domain interfaces as well as
the domains to evolve. To achieve the
full gain of the network architecture
transformation, the related internal
operator processes (like workflow, operation, and maintenance processes) will
need to be adjusted. Technologies like
SDN and cloud orchestration are crucial enablers and tools for automation
Conclusions
Increased levels of automation and
programmability are transforming
network architecture. This transformation is being driven by expected gains
in operational efficiency and reduced
TTM for new services, reduced TTC, and
new business, as well as by the fact that
enabling technologies such as virtualization and SDN are gaining maturity.
The target architecture is built on
interfaces that support the principles of
service and resource abstraction, multitenancy and programmability. Interdomain interfaces also support business
relations, as they include security and
SLAs, as well as separation of responsibility and accountability.
As a first important transformation
step toward the target architecture,
FIGURE 7 Architecture evolution
Business frame agreement
interfaces, non-real time,
not automated
Business and cross-domain
operations
Business agreement
interfaces, rather
static, not automated
BSS
Business management
X-dom
COMPA
Isolated per-tenant
EPC instances
Management
configuration
interface
COMPA
Expose
EPC
Expose
EPC
EPC
COMPA
Current architecture
E R I C S S O N R E V I E W • 2/2014
R&S
Process
Target architecture
Store
APIs within frame
agreement, real-time
and programmable
53
network programmability, but network
operations and services also need to be
controlled through operational policies
linked to business policies.
Due to the impact on operator processes and potentially even the business
ecosystem it is likely that the transformation will take place in a stepwise manner
over a significant period of time – with
different parts of the network evolving
at different rates. In addition, the resulting network architecture will support
5G radio evolution and the associated use
cases and requirements.
BOX B Main principles of the target architecture
Separation of concerns
Each domain has full responsibility over the resources and
operations performed inside the domain.
Exposure and abstraction of capabilities
The abstraction of functions into APIs that are exposed as
services supports domain inter-operability, which enables
automation and programmability.
Multi-tenancy
Each domain offers full isolation of how the different users
(tenants) use domain resources.
Intra-domain programmability
This is achieved by leveraging automation and
programmability within an administrative domain through its
COMPA functions.
Inter-domain programmability
Each domain exposes capabilities and services using welldefined APIs to achieve an end-to-end service offering,
orchestrated by the cross-domain COMPA functionality.
References
1. ETSI, 2014, Draft Group Specification, Security and Trust
Guidance, NFV ISG Spec, available at: http://docbox.
etsi.org/isg/nfv/open/Latest_Drafts/nfv-sec003v111
security and trust guidance.pdf
2. Ericsson Review, May 2014, IP-optical convergence: a
complete solution,available at: http://www.ericsson.com/
news/140528-er-ip-optical-convergence_244099437_c
3. Ericsson Review, February 2013, Software-defined
networking: the service provider perspective,
available at: http://www.ericsson.com/news/130221software-defined-networking-the-service-providerperspective_244129229_c
4. Ericsson Review, March 2014, Virtualizing network
services – the telecom cloud, available at: http://www.
ericsson.com/news/140328-virtualizing-networkservices-the-telecom-cloud_244099438_c
5. Ericsson Review, July 2014, Communications as a cloud
service: a new take on telecoms, available at: http://www.
ericsson.com/news/140722-communications-as-acloud-service-a-new-take-on-telecoms_244099436_c
6. Ericsson Review, June 2014, 5G Radio Access,
available at: http://www.ericsson.com/
news/140618-5g-radio-access_244099437_c
E R I C S S O N R E V I E W • 2/2014
Programmable networks
54
Torbjörn Cagenius
Göran Rune
Ignacio Mas
is an expert in distributed
network architecture at
Business Unit Cloud and IP. He
joined Ericsson in 1990 and has
worked in a variety of technology areas such
as FTTH, main-remote RBS, FMC, IPTV,
network architecture evolution, SDN and
NFV. In his current role, he focuses on cloud
impact on network architecture evolution.
He holds an M.Sc. from KTH Royal Institute
of Technology, Stockholm, Sweden.
is a principal researcher at
Ericsson Research. His current
focus is the functional and
deployment architecture of
future networks, primarily 5G. Before joining
Ericsson Research, he held a position as an
expert in mobile systems architecture at
Business Unit Networks focusing on the
end-to-end aspects of LTE/EPC, as well as
various systems and network architecture
topics. He joined Ericsson in 1989 and has
held various systems management
positions, working on most digital cellular
standards, including GSM, PDC, WCDMA,
HSPA, and LTE. From 1996 to 1999, he was a
product manager at Ericsson in Japan, first
for PDC and later for WCDMA. He was a key
member of the ETSI SMG2 UTRAN
Architecture Expert group and later 3GPP
TSG RAN WG3 from 1998 to 2001,
standardizing the WCDMA RAN
architecture. He studied at the Institute of
Technology at Linköping University,
Sweden, where he received an M. Sc. in
applied physics and electrical engineering
and a Lic. Eng. in solid state physics.
is a system architect at Group
Function Technology and an
expert in network architecture.
He holds a Ph.D. in
telecommunications from KTH Royal
Institute of Technology, Stockholm, and an
M.Sc. from both KTH and the Technical
University of Madrid (UPM). He joined
Ericsson in 2005 and has worked in IETF
standardization, IPTV and messaging
architectures, as well as media-related
activities for Ericsson Research. He is a
member of the Ericsson System Architect
Program (ESAP) and has research interests
in QoS, multimedia transport, signaling and
network security, IPTV and, most recently in
cloud computing.
Erik Westerberg
joined Ericsson from MIT,
Massachusetts, the US, in
1996 and currently holds the
senior expert position in
system and network architecture. In his
first 10 years at Ericsson, he worked with
the development of mobile broadband
systems before broadening his scope to
include the full network architecture,
serving as chief network architect until
2014. He holds a Ph.D. in quantum physics
from Stockholm University, Sweden.
Henrik Basilier
Lars Angelin
is an expert at Business Unit
Cloud and IP. He has worked for
Ericsson since 1991 in a wide
range of areas and roles. He is
currently engaged in internal R&D studies
and customer cooperation in the areas of
cloud, virtualization and SDN. He holds an
M.Sc. in computer science and technology
from the Institute of Technology at
Linköping University, Sweden.
is an expert in the multimedia
management technology area
at Business Unit Support
Solutions. He has more than 28
years of experience in the areas of concept
development, architecture and strategies
within telecom and education. He joined
Ericsson in 1996 as a research engineer, and
in 2003 he moved to the position of concept
developer for telco-near applications,
initiating and driving activities mostly
related to M2M and OSS/BSS. He holds an
M.Sc. in engineering physics and a Tech.
Licentiate in tele-traffic theory from Lund
Institute.
E R I C S S O N R E V I E W • 2/2014
Balázs Varga
joined Ericsson in 2010 and
he is an expert in multiservice
networks at Ericsson Research.
His focus is on packet evolution
studies to integrate IP, Ethernet and MPLS
technologies for converged mobile and fixed
network architectures. Prior to Ericsson, he
worked for Magyar Telekom on the
enhancement of broadband services
portfolio and introduction of new broadband
technologies. He has many years of
experience in fixed and mobile
telecommunication and also represents
Ericsson in standardization. He holds a
Ph.D. in telecommunication from the
Budapest University of Technology and
Economics, Hungary.
Acknowledgements
The authors gratefully acknowledge their
colleagues who have contributed to
this article: Jaume Rius i Riu and Ulf Olsson.
Ericsson
SE-164 83 Stockholm, Sweden
Phone: + 46 10 719 0000
ISSN 0014-0171
297 23-3237 | Uen
Edita Bobergs, Stockholm
© Ericsson AB 2014