Spectrum Decision in Cognitive Radio Networks - CSIR Research Space

14 downloads 1322 Views 6MB Size Report
modelling, which is commonly based on historical data to provide the means for ... As the demand for advanced broadband wireless technologies and services.
1

Spectrum Decision in Cognitive Radio Networks: A Survey Moshe Timothy Masonta, Student Member, IEEE, Mjumo Mzyece, Member, IEEE, and Ntsibane Ntlatlapa, Senior Member, IEEE

Abstract—Spectrum decision is the ability of a cognitive radio (CR) to select the best available spectrum band to satisfy secondary users’ (SUs’) quality of service (QoS) requirements, without causing harmful interference to licensed or primary users (PUs). Each CR performs spectrum sensing to identify the available spectrum bands and the spectrum decision process selects from these available bands for opportunistic use. Spectrum decision constitutes an important topic which has not been adequately explored in CR research. Spectrum decision involves spectrum characterization, spectrum selection and CR reconfiguration functions. After the available spectrum has been identified, the first step is to characterize it based not only on the current radio environment conditions, but also on the PU activities. The second step involves spectrum selection, whereby the most appropriate spectrum band is selected to satisfy SUs’ QoS requirements. Finally, the CR should be able to reconfigure its transmission parameters to allow communication on the selected band. Key to spectrum characterization is PU activity modelling, which is commonly based on historical data to provide the means for predicting future traffic patterns in a given spectrum band. This paper provides an up-to-date survey of spectrum decision in CR networks (CRNs) and addresses issues of spectrum characterization (including PU activity modelling), spectrum selection and CR reconfiguration. For each of these issues, we highlight key open research challenges. We also review practical implementations of spectrum decision in several CR platforms. Index Terms—Cognitive Radio, Primary User, Reconfiguration, Secondary User, Spectrum Characterization, Spectrum Decision, Spectrum Selection.

I. I NTRODUCTION

R

ECENT advancements in wireless technologies, such as software defined radios (SDRs), promise to address some of the major limitations experienced in legacy wireless communication systems. One of these limitations is inefficient utilization and management of the radio frequency (RF) spectrum in both licensed and unlicensed bands. Traditionally, RF spectrum is managed by the regulatory agencies through the assignment of fixed portions of spectrum to individual users in the form of renewable licenses. Although this regulatory Manuscript received 14 October 2011; 1st revision resubmitted March 2012; 2nd revision resubmitted 17 September 2012; manuscript accepted 24 October 2012. M. T. Masonta is with the Council for Scientific and Industrial Research (CSIR) and also a doctorate candidate at Tshwane University of Technology (TUT), Pretoria, Republic of South Africa (RSA), (e-mail: [email protected]), M. Mzyece is with the French South African Institute of Technology (F’SATI), Dept. of Electrical Engineering, TUT, Pretoria, RSA, (e-mail: [email protected]), and N. Ntlatlapa is also with the CSIR, Pretoria, RSA, (e-mail:[email protected]). The financial support of the CSIR is gratefully acknowledged.

approach ensures interference-free communications between radio terminals, it suffers from inefficient spectrum utilization. The available literature shows that spectrum utilization, on a block of licensed RF band, varies from 15% to 85% at different geographic locations at a given time [1]–[3]. As the demand for advanced broadband wireless technologies and services increases, traditional static spectrum regulation policies are becoming obsolete. To keep up with growing demand, there is a need for more efficient dynamic spectrum access (DSA) [4] technologies and regulatory approaches. The need for DSA or opportunistic spectrum access (OSA) was first proposed for the United States by the Federal Communications Commission (FCC) in 2003 [1]. This need was mainly driven by the threat of lack of operating spectrum for future wireless technologies. This move was recently followed by another important decision by the FCC in 2008 [5] and the Office of Communications (Ofcom) in the United Kingdom in 2010 [6], to open up television white spaces (TVWS) for unlicensed utilization. TVWS refers to large portions of RF spectrum, in the very high frequency (VHF) and ultra high frequency (UHF) bands, that will become vacant after the switch-over from analogue to digital TV [7]. Alongside these developments, there has been a strong trend towards research and development of cognitive radio (CR) [8] technology to optimally access the usable spectrum opportunistically and dynamically. A CR is “an intelligent wireless communication system capable of changing its transceiver parameters based on interaction with the external environment in which it operates” [1]. CR is therefore seen as an enabling technology for efficient DSA. While several approaches are proposed for achieving DSA (such as the dynamic exclusive use model, the spectrum commons model, and the hierarchical access model [4]), our focus in this paper shall be on DSA using CR technology. The process of realizing efficient spectrum utilization using CR technology requires a dynamic spectrum management framework (DSMF). In this paper, we shall adopt the DSMF proposed in [2] due to its clarity and relevance to our discussion. This DSMF consists of spectrum sensing, spectrum decision, spectrum sharing and spectrum mobility, as shown in Fig. 1. Spectrum sharing refers to coordinated access to the selected channel by the secondary users (SUs) or CR users. (While the terms “SU” and “CR user” are used interchangeably, in this paper we shall only use the term SU). Spectrum mobility is the ability of a CR to vacate the channel when a licensed user is detected. Spectrum sensing involves identification of spectrum holes and the ability to quickly detect the onset of

2

Fig. 1: Dynamic Spectrum Management Framework [2] Fig. 2: Spectrum Decision Framework licensed or primary user (PU) transmissions in the spectrum hole occupied by the SUs. Spectrum decision refers to the ability of the SUs to select the best available spectrum band to satisfy users’ quality of service (QoS) requirements. In this paper we will focus on the spectrum decision component of the DSMF. Spectrum decision involves three main functions [2]: spectrum characterization, spectrum selection and CR reconfiguration. Once vacant spectrum bands are identified (using spectrum sensing, geo-location databases or other techniques), each spectrum band is characterized based on local observations and on statistical information of the primary networks (which is normally called PU activities). The second step involves the selection of the most appropriate spectrum band, based on the spectrum band characterization. Thirdly, a CR should be able to reconfigure its transceiver parameters to support communication within the selected spectrum band. The required functions for the spectrum decision framework are summarised in Fig. 2. In order to perform these functions, the following questions need to be answered: 1. How can the available spectrum be characterized? 2. How can the best spectrum band be selected to satisfy the SU’s QoS requirements? 3. What is the optimal technique to reconfigure the CR for the selected spectrum band? (And how?) The above questions form the basis of spectrum characterization, spectrum selection and CR reconfiguration, respectively, as shown in Fig. 2. In this paper, we provide a comprehensive, up-to-date survey of the key research work on spectrum decision in cognitive radio networks (CRNs). We also identify and discuss some of the key open research challenges related to each aspect of the spectrum decision framework. This paper surveys the literature over the period 2003 to mid-2012 on spectrum decision in CRNs. This survey does not cover work done on spectrum sensing, spectrum sharing, spectrum mobility or geo-location databases. We choose to focus on spectrum decision because of its importance in and centrality to the DSMF in CRNs and because it has received relatively little attention compared to other components of the CR DSMF (namely spectrum sensing, spectrum mobility and spectrum sharing). In many ways, spectrum decision represents the culmination of the DSMF

in CRNs. We can limit our focus to this one aspect of DSMF based on the well-known communications engineering principles of modularity and abstraction, perhaps most famously and powerfully exemplified in Shannon’s 1948 classic paper [9]. The remainder of this paper is arranged as follows. Section II provides the background of CR technology and the motivation for performing spectrum decision in CRNs. Section III outlines CR standardization and regulation activities which are related to spectrum decision in CRNs. Section IV discusses spectrum characterization, the first of the three major spectrum decision functions in CRNs. Section V focuses on spectrum selection, the second major spectrum decision function in CRNs. Section VI covers CR reconfiguration and reconfigurable parameters, the final major spectrum decision function in CRNs. Section VII presents related work on practical implementations of spectrum decision on CR platforms. Future developments in CRNs are reviewed in Section VIII. Section IX concludes the paper. II. OVERVIEW OF CRN S In this section we provide an overview of CR technology and different CRN topologies. We briefly mention generic problems affecting spectrum decision functions due to the time-varying nature and fluctuations of the available spectrum in CRNs. A. Cognitive Radio Overview Recently, CR has received considerable attention from the research community as an enabling technology for efficient management of RF spectrum. In order to achieve DSA, a CR should be both spectrum and policy agile [10]. A spectrum agile CR is capable of operating over a wide range of frequency spectrum; while a policy agile CR will be aware of the constraints under which it operates (such as the rules for opportunistically using the vacant spectrum bands). Practically, CR builds on the software defined radio (SDR) architecture with added intelligence to learn from its operating environment and adapt to statistical variations in the input stimuli for efficient resource utilization [11]. With the current threat of

3

Fig. 3: Centralized CRN Topology

spectrum scarcity, CRs are widely proposed to build DSAbased secondary networks for lower priority users. One of the major functions of a CR is to find spectrum holes and be able to access and utilise them without causing any harmful interference to the incumbent or PU. A spectrum hole is defined as a band of frequency assigned to the PU, but which at a particular time and specific geographic location is not being used by that PU [11]. In the absence of signalling between PUs and SUs, spectrum holes may be identified by performing direct spectrum sensing, using geo-location databases, beaconing techniques, or by combining spectrum sensing with geo-location database information [7], [12]. (For interested readers the latest developments on database based CRNs, known as “SenseLess CRNs”, are reported in [13].) A CR should also be intelligent enough to perform spectrum decision in order to select the most suitable frequency band to satisfy specific communication needs. Spectrum decision is a key function of CRs which requires greater attention in order to realize the practical implementation and deployment of CRNs. Consequently, the focus of this paper will be on spectrum decision frameworks in both centralized and distributed CRNs. Readers are referred to [2] and [11] for good introductions to spectrum agile CR technology. B. Cognitive Radio Network Topologies A CRN is a wireless communication network whose enduser nodes are CRs. Similar to traditional wireless networks, a CRN topology can be classified as either centralized (infrastructure-based) or distributed (infrastructure-less or ad hoc) network topology. These network types are depicted in Fig. 3 and Fig. 4, respectively. In this paper, we consider both centralized and distributed CRN topologies. 1) Centralized CRN Topology In the infrastructure-based CRN architecture, a central node such as a base station (BS) or access point (AP) is deployed with several SUs associated with it, as shown in Fig. 3.

A typical example of a centralized CRN is a IEEE 802.22 wireless regional area network (WRAN) or a cellular network. For simplicity, we shall take the IEEE 802.22 WRAN as an example to discuss a centralized CRN. However, similar reasoning can be applied to more complex centralized CRNs. In a centralized network, a BS controls all the SUs (clients) or consumer premises equipments (CPEs) within its transmission range. The CRN operates within the transmission or coverage area of the primary network. Thus it uses DSA techniques to opportunistically access the primary network spectrum without causing any harmful interference. To do this, all SUs perform spectrum observation on specified spectrum channels and then send their observations to the BS, which acts as a fusion centre. Both the BS and its associated clients may be capable of detecting the presence of the PUs using different detection techniques (such as spectrum sensing, geo-location databases or beaconing). In some cases [14], two physical channels are used: one for observing the primary channel and the other for reporting data by the SUs to the BS. Once available channels are gathered, a BS will build the final list of these available channels and their associated maximum transmission powers, and then decide on the best channels to be accessed. These channels will then be broadcast back to all or selected SUs for use. In the next subsection we discuss the distributed CRN topology. 2) Distributed CRN Topology In the distributed CR ad hoc network (CRAHN) topology, the SUs communicate directly with each other without any central or controlling node. As shown in Fig. 4, SUs share their local observations and analysis among themselves, as long as they are within each other’s transmission range. For database-based networks, each SU may have access to query the database for available spectrum bands. Using both its results and the results of other SUs, a SU can make a decision for an appropriate band using a local criterion. If the criterion is not satisfied, the process may be repeated again until a decision is reached. It is clear that spectrum decision in CRAHNs does not rely on a central node. However, if SUs decide to cooperate, as in cooperative spectrum sensing, one node can be chosen as the head node and be used for making spectrum decisions. Unlike in infrastructure-based topologies, spectrum decision in CRAHNs also involves route selection, which is normally addressed as a joint spectrum and route selection problem. A noticeable new challenge in CRAHNs, which did not exist in traditional wireless ad hoc networks, is that channel availability is determined by the present behaviour of PUs, which may vary with location, time and frequency [15]. Another new challenge is the re-routing and switching to other available channels or links once the PU appears on the occupied channel [16]. Thus the wide range of operating or available spectrum makes it infeasible to transmit beacons over all possible channels. Section V discusses these and many other spectrum selection challenges experienced in both distributed and centralized CRNs. In this section, we have provided a brief overview of

4

Fig. 4: Distributed CRN Topology

CR technology in relation to spectrum decision and the two most commonly deployed CRN topologies (i.e. centralized and distributed topologies). In the next section, we present standardization and regulation efforts around CR technology. III. S TANDARDIZATION AND R EGULATORY E FFORTS The introduction of frequency agile CR technology created a spark in numerous academic, industry, regulatory and standardization bodies worldwide. Like any other new technology introduced to the market, CR technology’s success will depend on sound standardization and regulation efforts from standardization bodies, regulators and industry. It is important to note that initial research and development efforts on CR technology have been focused in the United States. This is mainly due to the FCC’s adoption of CR as an enabling technology for efficient spectrum management. Since then, other standardization bodies and regulatory agencies around the world have become interested in the standardization and regulation of CRs for DSA. In the following sub-sections we will discuss standardization and regulatory efforts on CRs, focusing mainly on efficient frequency management. A. Standardization 1) IEEE 802.22 Standard The global switch-over from analogue to digital TV will leave a considerable amount of VHF/UHF spectrum vacant. The TVWS spectrum has excellent radio propagation characteristics, and is now being proposed as the most useful spectrum for improving wireless broadband connectivity in rural communities [12], [17]. In order to take advantage of the TVWS spectrum, the IEEE 802.22 WRAN standard [18] was established, and the first official standard was released in July 2011. IEEE 802.22 is the first wireless air interface standard focused on the development of CR based WRAN physical (PHY) and medium access control (MAC) layers for operation in TVWS. It specifies a fixed point-to-multipoint

wireless air interface where a BS manages its own cell and all associated CPEs. The IEEE 802.22 PHY layer is based on orthogonal frequency division multiple access (OFDMA) and can support a system which uses TVWS channels to provide wireless communication links over distances of up to 100 km. A typical use case for the IEEE 802.22 standard would be in sparsely populated rural areas [7]. The IEEE 802.22 standard supports incumbent or PU detection through spectrum sensing techniques with an option for geo-location databases. However, there are still technical difficulties in performing reliable spectrum sensing practically. Thus, some regulatory bodies such as the FCC and Ofcom prefer geo-location databases as the primary means for incumbent detection. A beaconing option is also provided for incumbent user detection in IEEE 802.22. In IEEE 802.22, both the CPE and BS have the capability to detect the incumbent, but spectrum decision is only managed by the central BS. The BS employs the CR capabilities for spectrum decision based on the TV channels’ operating characteristics [19]. The BS actually performs the spectrum characterization and selection functions, while the CPE is responsible for the reconfiguration of its transceiver parameters. 2) IEEE DySPAN After realizing the importance of coordinated work around CR standardization, the IEEE P1900 Standards Committee was jointly established by the IEEE Communications Society (ComSoc) and the IEEE Electromagnetic Compatibility Society in the first quarter of 2005 [20]. On 22 March 2007, the IEEE Standards Association Standards Board approved the reorganization of the IEEE P1900 activities as Standards Coordinating Committee 41 (SCC41), called Dynamic Spectrum Access Networks (DySPAN). The main aim of SCC41 is to develop supporting standards to address issues related to new technologies and the development of techniques for next generation radio systems and advanced spectrum management [21]. The SCC41 concentrates on developing architectural concepts and specifications for network management between incompatible wireless networks rather than specific mechanisms that can be added to the air interface. In December 2010, the IEEE SCC41 was renamed the IEEE DySPAN-Standard Committee (DySPAN-SC). The IEEE DySPAN-SC consists of seven working groups (WGs), named 1900.1 through to 1900.7. Out of these WGs, the IEEE 1900.4’s work has some elements of spectrum decision. This WG focuses on architectural building blocks enabling network-devices decision making for optimized radio resource usage in heterogeneous wireless access networks [20]. 3) European Telecommunications Standards Institute In Europe, the European Telecommunications Standards Institute (ETSI) is also involved in the standardization of CR systems (called reconfigurable radio systems) under their Reconfigurable Radio Systems Technical Committee (RRSTC) [22]. Cognitive radio principles within ETSI RRS-TC are concentrated on two topics: a cognitive pilot channel proposal and a functional architecture for management and control of reconfigurable radio systems. There are four WGs forming

5

the ETSI RRS-TC, WG 1 to WG 4. Cognitive management and control falls under WG 3. This WG focuses on defining the system functionalities for reconfigurable and dynamic spectrum management and joint radio resource management. More information on ETSI RRS-TC can be found in [22].

B. Regulation The International Telecommunication Union (ITU) is also involved in standardization efforts of CR technology through their ITU-R Working Party (WP) 1B and WP 5A [23]. These two WPs prepared reports describing the concepts and the regulatory measures required to introduce CR. The ITU-R WP 1B developed a working document towards draft text on World Radio-communications Conference 2012 (WRC-12) agenda item 1.19. Agenda item 1.19 reads: ”to consider regulatory measures and their relevance, in order to enable the introduction of software-defined radio and CR systems, based on the results of ITU-R studies, in accordance with Resolution 956 of WRC 07” [24]. The ITU-R WP 5A is currently developing the working document toward a preliminary new draft report, Cognitive Radio Systems in the Land Mobile Service [25]. This report will address the definition, description, and application of CR systems in the land mobile service [23]. The regulatory technicalities on dynamic spectrum management and spectrum decision from the ITU’s point of view should become clearer after the WRC-12, where they will be discussed under agenda item 1.19. Now that we have presented the necessary background to assist in understanding CR technology, CRN topologies, and CRN-related standardization and regulation activities, in the following sections we will discuss the three major functions in the spectrum decision framework. These will be followed by additional sections on spectrum decision in CR platforms and future developments in CR technology. IV. S PECTRUM C HARACTERIZATION IN CRN S In CRNs, multiple spectrum bands with different channel characteristics may be found to be available over a wide frequency range [26]. In order to properly determine the most suitable spectrum band, it is crucial to first identify the characteristics of each available spectrum band. Spectrum characterization allows the SUs to characterize the spectrum bands by considering the received signal strength, interference and the number of users currently residing in the spectrum, based on RF observation. The SUs should also observe heterogeneous spectrum availability which varies over time and space due to PU activities. Heterogeneous spectrum availability refers to the availability of spectrum holes which fluctuate over time and location and have different characteristics. Thus, spectrum characterization should include both the current RF environment conditions and the observed PU activity modelling. In this section, spectrum characterization in terms of the radio environment and PU activity models is discussed along with some related work. The section ends with key open research challenges in spectrum characterization.

Fig. 5: Radio frequency environment characterization elements

A. Radio Frequency Environment Characterization A CR is expected to continuously characterize radio environment usage in frequency, time and space. This is mainly due to the fact that available spectrum bands in CRNs always have different characteristics. Radio frequency (RF) environment characterization is a process that involves estimation of the following key elements or parameters: (1) channel identification, (2) channel capacity, (3) spectrum switching delay, (4) channel interference, (5) channel holding time (CHT), (6) channel error rate, (7) subscriber location, and (8) path loss [2], [27]. These elements are illustrated in Fig. 5 and analysed and discussed in the following subsections. 1) Channel Identification Primary channel identification is the first important step to be performed by each CRN. As research in CR evolves, different CRNs application areas are being introduced to the market. Some of these applications include: television (TV) white space networks, smart grid networks, machine-tomachine (M2M) networks, public safety networks, broadband cellular networks and wireless medical networks [28]. These applications or networks exhibit different traffic data patterns, either deterministic or stochastic. In deterministic traffic, the PU is assigned a fixed time slot on a frequency band for communication. Once the PU stops communicating, the frequency band becomes available and can be used by the SUs. Examples of deterministic traffic data patterns occur in TV broadcasting (longer periods) and radar transmitters (shorter periods). Normally deterministic signals have fixed or predictable ON and OFF periods which can be determined by a mathematical expression, rule or table. Any future value for a deterministic signal can be calculated or predicted based on its past values, which makes it easy to predict future PU idle periods for CRNs. On the other hand, stochastic traffic patterns can only be described and analysed using probabilities and statistics because their spectrum usage tends to exhibit greater variations in time and space [11]. Due to their randomness, stochastic signals are analysed using average values from a collection of primary signals. Examples of stochastic traffic data patterns occur in cellular networks. In order to improve the accuracy of stochastic traffic modelling, Haykin [11] suggested the design

6

of a tracking strategy for PU idle period prediction models. A prediction method for both deterministic and stochastic traffic patterns is proposed in [29]. This method is used by CRs to predict the primary channel idle time. In summary, channel identification is a crucial function for every CRN for learning its external environment, classifying primary traffic and applying the appropriate spectrum decision methods. 2) Channel Capacity Estimation While many other parameters (such as the channel interference level, error rate, path loss, delay and holding time) are important for efficient spectrum characterization, a considerable amount of research has focused on channel capacity estimation. This is motivated by the fact that by estimating other (above) channel parameters, we can determine the channel capacity [2] (this means channel capacity can be derived from the above parameters). It has been shown that the traditional method of estimating channel capacity using the signal-to-noise-ratio (SNR) leads to non-optimal spectrum decision [30]. In an orthogonal frequency division multiplex (OFDM) system, each spectrum band i has a different bandwidth Bi , consisting of multiple subcarriers. A normalized CR capacity CiCR (k) model of spectrum band i for user k is proposed in [26] for spectrum characterization in CRNs. This CiCR model defines the expected normalized capacity of user k in spectrum band i as:

CiCR (k) = E[Ci (k)] =

Tiof f of f Ti +

τ

.γi .ci (k)

(1)

where Ci (k) represents the spectrum capacity, the term ci (k) (with small c) is the normalized channel capacity of spectrum band i in bits/sec/Hz, τ represents the spectrum switching delay, γi represents the spectrum sensing efficiency and Tiof f is the expected transmission time without switching in spectrum band i. Spectrum or channel switching delay is introduced within CRNs when SUs move from one spectrum band to another according to PU activity. More on channel switching delay is discussed in the next sub-section. Spectrum sensing efficiency arises due to the fact that RF front-ends cannot perform sensing and transmission at the same time, which inevitably decreases their transmission opportunities. When using spectrum sensing to detect spectrum holes, sensing efficiency is influenced by the observation time and transmission time [31]. 3) Channel Switching Delay In opportunistic or DSA network settings, radio nodes are expected to operate on different frequency channels in a given time without disrupting existing network connections [32]. This process is commonly known as dynamic channel switching (DCS). In CRNs, channel switching may be triggered by the detection of PUs on the operating channel, by degradation of the QoS due to interference, or traffic load in the current channel [18]. During the channel switching process, the SUs must dynamically switch from one channel

to another idle channel, and during this switching process all SUs transmissions are temporarily suspended until the new spectrum opportunity or channel is found. Switching from one channel to another introduces additional delay to the CRNs, which is called switching delay. This switching delay may vary from one node to another since it depends on the hardware technology (e.g. time spent during RF front-end reconfiguration) and the algorithm used by the SUs or the overall CRNs to perform the spectrum decision process. The switching delay may also include the sensing time in cases where available channels are detected using spectrum sensing techniques. The ultimate goal is to keep the switching delay as short as possible to ensure that it does not affects the overall CRN performance. Azarfar et al. [33] addresses the channel switching delay as an overhead that has the potential to decrease the useful time available for data communication. They consider channel switching delay as the sum of spectrum sensing duration plus the channel recovery time. Channel recovery time is the time spent by a SU to vacate the channel, to decide on the available channel (or signalling time for establishing new channels), and to select or access the available spectrum. Several techniques are proposed in [33] to reduce the channel switching delay. This includes the use of historical information of channel occupancy and channel quality index [33]. Historical information of channel occupancy becomes useful provided the SUs are aware of their operating location. For instance, a SU may use geographical coordinates of the previously explored areas to remember that spectrum band X was occupied or vacant and also to remember which technologies operate in that area. So when the SU approaches those areas (i.e. previously explored areas), it can save time by avoiding those spectrum bands which were found to be occupied during the previous visits. This will then reduce the channel recovery time, thereby reducing the channel switching delay. In centralized CRNs such as IEEE 802.22 [18], the BS can decides to switch channels during normal operations by first selecting the backup channel from the backup or candidate channel list. Secondly, it must wait for a specified time to make sure that all the associated CPEs are prepared for the channel switch. This waiting time should be long enough for the CPEs to recover from an incumbent detection. Once all CPEs are prepared for channel switching, the BS can then schedule the channel switching procedure. Typical spectrum switching delay in IEEE 802.22 is less than 2 seconds [18]. Xu et al. [34] analysed a trade-off between higher bandwidth and switching overhead experienced in CRNs. In their paper, switching overhead is the sum of spectrum sensing time, channel evacuation time, and link setup time. The link setup and channel evacuation times are based on the radio hardware and the operating environment [34], and are modelled as a random variable for all SUs. It was found that using higher bandwidth in saturated traffic (i.e. where there is a large pool of idle PU channels to be sensed) with few SUs leads to more switching overhead due to an increased sensing time. However, in cases where a central node provides channel availability information to multiple SUs (i.e. SUs do not need to conduct spectrum sensing), the switching overhead reduces (it only

7

consists of link setup and channel evacuation times). In distributed CRNs, channel switching overhead on a node includes the switching delay, other flows’ transmissions delay and the back-off delay (i.e. interference within a frequency channel) [35]. A delay based routing metric called cumulative delay for on-demand routing protocol is proposed in [35]. This cumulative delay is the total path delay derived from the spectrum switching delay and back-off delay. The backoff delay arises when more than two nodes contend for the spectrum resource. This delay will depend on the number of contending nodes on each spectrum band. Thus during spectrum characterization, the number of SUs in a distributed CRNs is important to estimate the switching delay. An analysis of delay performance in CR sensor networks is presented in [36]. Two types of channel switching techniques are considered: periodic switching and triggered switching. In periodic switching, SUs switch to a new channel only at the beginning of each channel switching interval as defined in [36]. Periodic switching occurs if the occupied channel becomes unavailable (to the SUs) before the end of the channel switching interval. Channel switching can be caused by the sudden appearance of PUs or high interference on the channel due to adjacent channels. In triggered switching, the SUs switch to a new channel as soon as the current channel is no longer available [36]. It was found that bursty traffic experiences shorter delays when considering periodic switching techniques as compared to triggered switching techniques. This was mainly due to the fact that a channel is likely to be available in the earlier portion of the channel switching interval than in the later portion. 4) Channel Interference Estimation In CRNs environment, SUs are expected to coexist with licensed or primary users (PUs). In some cases, several CRNs must also coexist within the coverage area of single or multiple primary networks. Such coexistence, if not controlled, can lead to harmful interference to the PUs. It is therefore crucial to accurately estimate and model interference generated by multiple active SUs in the network. In this subsection, different techniques for estimating, controlling and modelling channel interference caused by CRNs are discussed. In [37], an opportunistic interference alignment scheme that allows multiple SUs to exploit the unused spatial dimensions of multiple-input-multiple-output (MIMO) PU channels is proposed. In this scheme, the primary transmitters maximize their rate by water-filling over the singular values of their channel matrix. By using singular values, PUs leave some eigen-modes unused, which allows the SUs to transmit at a significant rate by aligning their signals along the free eigenmodes of the PU channel. As a result, this scheme protects the transmission of the PUs while providing interference-free communication for the SUs. An interference-aware radio resource allocation scheme is proposed in [38]. The paper studies PU interference caused as a result of: CR out-of-band (OOB) emissions and the interference that arises as a result of imperfect spectrum sensing. In OFDM-based primary networks, OOB emissions are due to power leakage in the side-lobes of transmitted

signal. The amount of OOB interference power introduced in a PU sub-carrier due to SU transmission is modelled for both the uplink and downlink sub-bands. Finally, a computationally efficient algorithm for downlink and uplink subcarriers and power allocation in an OFDMA-based CRN is developed based on proposed OOB emissions and imperfect-spectrumsensing-based interference models. To minimize the amount of OOB interference generated via non-contiguous multicarrier data transmission, multi-rate filter banks are suggested in [39]. In this approach, the multi-rate filter banks’ sub-band spectra can be designed to be highly spectrally selective to limit the amount of intercarrier interference (ICI), which becomes advantageous in cellular systems that suffer from Doppler effects and frequency selectivity of the channels. In [40], two interference detection schemes are proposed. The first scheme is based on pilot-aided interference detection for OFDM systems. In order to detect the presence of interference, this scheme requires at least two pilot symbols in a given subcarrier spaced in time. The pilot symbols are designed to ensure that their summation or subtraction is zero. For instance, two pilot symbols, xp1 and xp2 , can be selected as the two points in the Binary Phase Shift Keying (BPSK) constellation such that their summation is zero (i.e. xp1 + xp2 = 0). Although this pilot-aided scheme is simple in implementation, its weakness is poor interference detection in the sub-carriers where no pilot exists due to sparse placement of the pilot symbols. The second scheme proposed in [40] is based on a joint interference detection and decoding technique which does not require any pilot symbols. The decoder jointly performs erasure marking and decoding in order to erase the interference jammed symbols automatically during the decoding process. The decoding process consist of two steps: the first step determines the positions of the erasures, and the second step determines the number of erasures. However, this technique suffers from increased computational complexity and decoding delay. Rabbachin et al. [41] proposed a statistical model for perdimension (real or imaginary part) aggregate interference of a CRN which allows modelling of the CRN interference generated by SUs in a limited or finite region. In this model, two types of secondary spatial reuse protocols are considered: single-threshold and multiple-threshold protocols. For each protocol, the characteristic function of the CRN interference is expressed, and then used to derive its cumulants. The cumulants are used to model the CRN interference as truncatedstable random variables. The interference signal at the primary receiver generated by the ith cognitive interferer is modelled as:

Ii =

p

PI Ri−b Xi

(2)

where PI is the interference signal power at the limit of the near-far region (which is limited to 1m), Ri is the distance between the ith cognitive interferer and the primary receiver, b is the amplitude path-loss exponent, and Xi is the per-dimension fading channel path gain of the channel from the ith cognitive interferer to the primary receiver.

8

(CSMA) and OFDMA techniques. 5) Channel Holding Time Channel holding time (CHT) is the expected duration the SUs can occupy a licensed band before getting interrupted. Thus, the longer the holding time, the better the QoS for the SUs [2]. CHT can be determined by the type of secondary services served by the CRN or it can be determined by the regulator. It is useful in determining the RF environment characterization. In [42], a Markovian model for finding the duration of the spectrum hole is proposed. This model builds on a CHT concept for the PU. Once the PU idle time is modelled as CHT, matrix-analytic techniques are applied to derive and analyse the duration of the spectrum holes which can be accessed by the SUs. One of the main drawbacks of this technique lies in its complexity. Yuan et al. [43] introduced the concept of time-spectrum block to model spectrum reservation for CRNs. A timespectrum block concept represents the time for which a SU occupies a portion of vacant spectrum without causing interference to the PUs. For a CRN of n nodes (V = v1 , · · · , vn ), located in the two-dimensional Euclidean plane, let d(vi , vj ) denote the Euclidean distance between k vi and vj . A time-spectrum block Bij = (tk , 4tk , fk , 4fk ) is assigned to link (vi , vj ) if sender vi is assigned the contiguous frequency band [fk , fk + 4fk ] of bandwidth 4fk during time interval [tk , tk + 4tk ]. Using the above definitions, the dynamic spectrum allocation problem can be viewed as dynamic packing of time-spectrum blocks into a three-dimensional resource, consisting of time, frequency, and k space. Using a time-spectrum block Bij , it is possible to find the time overhead of switching frequency or the time used for medium access contention in CRNs. 6) Channel Error Rate In a communication link, error rate is defined as the rate of bits or data elements which are incorrectly received from the total number of bits or data elements sent during a specified time interval [44]. The average channel error rate is a useful parameter in estimating the RF environment characterization in CRNs. It depends on the interference level (interference to SU may be caused by the primary transmitters or other SUs), the available bandwidth, the frequency band in use and the modulation scheme (or access technology) [2]. Bit error rate (BER) and frame error rate (FER) are the most commonly used metrics. Error rate is usually stated relative to the channel’s signal to noise ratio (SNR) values, and this makes the transmitted energy per bit an important metric in error estimation [44]. A closed-form average BER expression is derived in [45] to investigate SU’s error performance for the binary phase shift keying (BPSK) modulation scheme. It was found that the channel error performance improves when the SU’s SNR increases. Kaur and Sharma [46] analysed the bit error rate (BER) performance of the CR PHY layer over Rayleigh fading channels under different channel encoding schemes and channel conditions. They considered CRNs with both contention-based non-persistent carrier sense multiple access

7) Subscriber Location Subscriber location also need to be determined when characterizing the RF environment. Generally, a SU can obtain geographical and environmental information using built-in global positioning system (GPS) coordinates, embedded information in packets exchanged between nodes or a central server that sends the most up-to-date global Radio Environment Map (REM) information [33]. IEEE 802.22 defines two modes which the SUs and BSs can use to find their geo-location: satellite-based geo-location (which is mandatory) and terrestrial-based geo-location (which is assisted by the CDMA ranging) [47]. By knowing its location, a SU can record a number of normal and abnormal events experienced in different locations and times. Such knowledge will be useful for future predictions of spectrum holes and characterization of the RF environment. 8) Path Loss Path loss, the deterministic overall reduction of received signal power with distance between the transmitter and the receiver, is one of the factors affecting the radio propagation across wireless channels [11]. It is caused by the spreading of the electromagnetic wave radiating from the transmit antenna and the obstructive effects of the surrounding objects [44]. Path loss is normally obtained at the receiver side by dividing the transmitted power with the received power. Thus, transmission power can be increased to compensate for the increased path loss. However, this might cause higher interference for other SUs and PUs. According to Rappaport [48], the average path loss of a channel can be expressed using a path loss exponent (α). This path loss exponent α depends on frequency, antenna heights, and propagation environment. In free space loss, the path loss exponent α is equal to 2. In [49], a low-complexity adaptive transmission protocol for CRNs, whose links have unknown and time-varying propagation losses, is proposed. This protocol adjusts its transmission modulation and coding as a mechanism for responding to changes in propagation losses. The transmitter power is increased only if the most powerful combination of coding and modulation is inadequate. In this subsection we reviewed related work on RF environment characterization, which is the first step towards reliable spectrum decision scheme development. In the next subsection, we discuss PU activity modelling as another aspect of spectrum characterization. B. Primary User Activity Modelling Because there is no guarantee that a spectrum band will be available during the entire SU communication period, it is important to consider how often the PUs appear on the spectrum band. Using the learning ability of the CR, the history of the spectrum usage information can be used for predicting the future profile of the spectrum. This process is achieved through PU activity modelling. By considering the PU activity, the SUs can decide on the best available

9

spectrum bands to be used for their transmissions. For a stable CRN, the spectrum decision framework must be aware of the available spectrum fluctuations, as well as the heterogeneous QoS requirements of the SUs [2]. It is important for the spectrum decision function to also consider the spectrum fluctuations because the SUs can transmit data only if they can accurately detect the spectrum holes. PU activity modelling also plays a crucial role in the design of communication protocols for CRNs [50]. For example: once we know that a PU favours a particular channel and tends to occupy it for a long period of time, that channel would be less likely to be available for a SU, as a result, sensing on such a channel would likely be a waste of time and energy [51]. Therefore, PU activity modelling will result in more effective spectrum usage for SUs, which in turn enhances CRNs performance [52]. Different techniques are used to model PU activities, and we discuss them in the next sub-sections. 1) PU Activity based on Poisson Modelling There is a significant amount of research that models the PU activity as a Poisson process with exponentially distributed inter-arrivals [26], [53], [54]. The PU traffic is modelled as a two-state birth-death process with death rate α and birth rate β. In this approach, each user arrival is independent, and the PU transmission is assumed to follow the Poisson arrival process. As a result, the length of ON and OFF periods are exponentially distributed. An ON state represents the period used by PUs and an OFF state represents the vacant periods. An adaptive spectrum decision framework is proposed in [26] to determine a set of spectrum bands to satisfy the user requirements under the dynamic nature of RF spectrum bands. In this framework, each spectrum band is first characterized by jointly considering PU activity and spectrum sensing results. For real-time applications, a minimum variance-based spectrum decision which minimizes the capacity variance of the decided spectrum bands is proposed. For best-effort applications, a maximum capacity-based spectrum decision scheme is proposed to maximize the total network capacity. However, based on a large-scale measurement-driven characterization of the PU activity in cellular networks investigated in [55], it was found that PU activity durations are non-exponential and high fluctuations of the PU activity may violate the Poisson assumption. While the majority of studies assume that the PU activity follows the Poisson model, such assumptions were invalidated in [52], [56] based on the following reasons: • The Poisson model approximates the PU activities as smooth and burst-free traffic. Therefore it fails to capture the bursty and spiky characteristics of the monitored data. • The Poisson model does not consider correlations and similarities within data. • The Poisson model fails to capture the short-term temporal fluctuations or variations exhibited by the PU activity. Therefore the existing research which assumes the Poisson model derives PU activity models that have smooth and burstfree traffic in which short-term fluctuations are neglected [52]. A good illustration of the Poisson model’s drawbacks can be found in [52], as depicted in Fig. 6. The Poisson model

Fig. 6: Missed Opportunities in Poisson Model [52] . represents the ON period (which is the active transmission duration of a PU) and the OFF period (which represents the absence of PU activity). It can be observed from Fig. 6 that the actual PU activity fluctuates during the ON period, which is not tracked by the Poisson model. As a result, these durations are classified as part of the ON period, which leads to missed transmission opportunities. Based on these reasons, it is clear that Poisson modelling may lead to performance degradation in CRNs due to unidentified fluctuations in PU activities. In order to address the potential Poisson model drawbacks, the following subsections presents alternative mechanisms for PU activity modelling. 2) PU Modelling Based on Statistics A simple method for learning and classifying traffic patterns on primary channels is proposed in [29]. The authors consider a primary network that consists of multiple channels with independent traffic patterns. The method starts by collecting spectrum usage information (through spectrum sensing) and stores this information in the channel database (in binary format). If the channel is free, the channel state (CS) flag is set to 0, and CS is set to 1 if the channel is occupied. Once the spectrum information is stored, the traffic patterns of each channel are classified as either stochastic or deterministic. The channel classification algorithm used is based on the periodicity, where an edge detection method is used for period search. The average separation of the raising edges and the standard deviation of the separations (of the edge) are used to determine whether the traffic is deterministic or stochastic. After classifying traffic type, the idle time prediction method is selected. To predict the PU idle time, the CS information (stored in the channel database) is used. If CS = 0, the idle time prediction is set to 1. If CS = 1, the idle time prediction is set to 0. Secondary transmission will continue on channels with longer idle time, and once the PU appears on this channel, the SU will switch to the channel with the longest expected remaining idle time. Fig 7 summarizes the model as proposed

10

by popular wireless services such as Wi-Fi, TV and cellular networks. However, it was found that Geary’s C requires sufficient amount of data to ensure that existing similarity patterns are not missed when the investigated bands are only rarely (i.e. in less busy locations/times). A time-varying statistical model for spectrum occupancy using actual wireless frequency measurements is proposed in [62]. Using statistical characteristics extracted from actual RF measurements, first-order and second-order parameters are employed in a statistical spectrum occupancy model based on a combination of different probability density functions.

Fig. 7: Prediction System Model [29]

in [29]. Acharya et al. [57] proposed a predictive model based on long-term and short-term usage statistics of TV channels. In this model, the usability characteristics of a channel are based on TV channel statistics and are later used for selection of a channel for opportunistic transmission. The model uses a threshold mechanism to filter out channels with frequent and heavy appearance of PUs. In [58], a multivariate time series approach is used to learn the PU characteristics and then predict the future occupancy of neighbouring channels. In order to reduce the complexity and storage requirements, a binary scheme is used; where 0 represents empty and 1 represents occupied channels. 3) PU Modelling Based on Measured Data Riihijarvi et al. [59] proposed a technique to characterize and model the spectrum for DSA with spatial statistics and random fields using measured data. The proposed techniques were implemented using data obtained through spectrum measurements. First, they perform spectrum characterization by treating spectrum measurement results as a realization of some unknown random field Z. In this case, random fields are considered as extensions of the theory of stochastic processes from one-dimensional time to multi-dimensional space. Different metrics suitable for characterizing second-order stochastic spatial structure of the primary spectrum are derived using Moran’s I [60] and Geary’s C [61] statistics. These are two statistics commonly used to measure the degree of spatial correlation for spatial data whose covariance structure is defined by neighbourhoods. Moran’s I statistics is defined as:

I=

n . S0

PP

¯ ¯ Wij (Zi − Z)(Z j − Z) P ¯ 2 (Zi − Z)

(3)

where n is the sample size, Wij is a matrix of weights (where Wij = P 1 P if site i, j are neighbours, otherwise Wij = 0). S0 = Wij = twice the number of neighbours, Zi indicates the continuous response at site i, and finally Z¯ denotes the usual estimate for the mean of Z. The value of Moran’s I lies between -1 and +1, where former indicating strong negative autocorrelation and latter strong positive autocorrelation. See [59] for Geary’s C statistics representation. The above spatial statistics models were tested on distributed spectrum occupancy measurements of frequency bands used

4) Other PU Modelling Techniques Canberk et al. [52] developed a real-time based PU activity model for CRNs using first-difference filter clustering and correlation. In this difference model, the PU signal samples are first collected over a pre-determined duration, then the observed PU signals are clustered together if they are greater than a threshold. The authors developed a PU activity monitoring module which is implemented within each SU to monitor the spectrum bands and then samples the PU activity. In simple terms, this module is responsible for local spectrum sensing by testing the well known H1 and H0 hypotheses [2], where H1 indicates that the PU has an activity in the spectrum band and H0 indicates that there is no PU activity in the spectrum band. This PU activity monitoring module stores samples of the monitored PU activities into a vector, q, of size p as [52]:

q = [q(1), q(2), · · · , q(m), · · · , q(p)] ∈