About this Journal Submit a Manuscript Table of Contents
The Scientific World Journal
Volume 2015 (2015), Article ID 394956, 16 pages
http://dx.doi.org/10.1155/2015/394956
Research Article

Reliable Adaptive Video Streaming Driven by Perceptual Semantics for Situational Awareness

Department of Telecommunications and Systems Engineering, Autonomous University of Barcelona, 08193 Bellaterra, Spain

Received 31 October 2014; Revised 1 April 2015; Accepted 21 April 2015

Academic Editor: Hsiang-Fu Yu

Copyright © 2015 M. A. Pimentel-Niño et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

A novel cross-layer optimized video adaptation driven by perceptual semantics is presented. The design target is streamed live video to enhance situational awareness in challenging communications conditions. Conventional solutions for recreational applications are inadequate and novel quality of experience (QoE) framework is proposed which allows fully controlled adaptation and enables perceptual semantic feedback. The framework relies on temporal/spatial abstraction for video applications serving beyond recreational purposes. An underlying cross-layer optimization technique takes into account feedback on network congestion (time) and erasures (space) to best distribute available (scarce) bandwidth. Systematic random linear network coding (SRNC) adds reliability while preserving perceptual semantics. Objective metrics of the perceptual features in QoE show homogeneous high performance when using the proposed scheme. Finally, the proposed scheme is in line with content-aware trends, by complying with information-centric-networking philosophy and architecture.

1. Introduction

Video services have become part of everyday interactions and contribute to a major portion of network traffic. Usage of video can surpass the recreational arena to provide additional insights in out-of-the-ordinary scenarios, such as emergencies or e-health aid.

Our interest is the use of live, point-to-point, beyond recreational video streaming. The goal of such service is to provide valuable information through the live video, in order to enhance the end-user’s awareness of ongoing situations. We consider scenarios where best effort satellite networks become a reliable alternative to unavailable terrestrial communications infrastructure. However, these networks offer stringent conditions that challenge the user’s satisfaction. The relevance of this realistic scenario is supported by direct contact and fieldwork with decision makers using these types of video services during disaster and emergency response events [1].

In this paper, we propose a novel, complete model and solution for live video transmission of user generated to enhance situational awareness. It is composed of novel QoE framework that allows fully controlled adaptation and enables perceptual semantic feedback. The framework is based on temporal/spatial abstraction for video applications serving beyond recreational purposes. To the best of our knowledge, the aim of offering complete solution for the given scenario requirements (video for beyond recreational needs to enhance awareness) and its constraints (challenging communications with scarce bandwidth suffering from both congestion and erasures) has not been addressed before. Moreover, we use novel tools to tackle these issues and provide robust and feasible scheme.

Existing Strategies for Congestion Avoidance. Video streaming is growing in popularity on the best effort internet. Existing technologies such as dynamic adaptive streaming over HTTP (DASH), layered on top of TCP, have well-known advantages. However, they are not applicable in our design for situation awareness where, for example, satcom scenarios need to be considered. In such scenarios, TCP has low performance for streaming due to long round trip times and packet erasures. This results in a decrease in throughput that causes long start-up delays in video playback and bounds the video source rate to just half of the TCP throughput [2]. At the same time, the topology of our scenario is not compatible with HTTP/TCP—HTTP options of user generated live (close to real-time) streaming rely on intermediate servers to prepare content for the clients (e.g., in order to use DASH). All of the above affect QoE with freezes in playback and low image quality. As an alternative, improvements to general purpose TCP for satellite communications include performance enhancement proxy (PEP) solutions; however, they alter the system architecture.

A variety of TCP-friendly congestion control schemes have been studied for video streaming [35]. Their main focus is the fairness of the schemes rather than the impact on QoE. Further, they rely on heavy feedback from the receiver, which is a problematic issue in long-delayed networks. Real-time applications, on the other hand, opt for real-time protocol (RTP) and use ad hoc congestion control schemes [6, 7]. They offer the flexibility our scenario needs at the transport layer; however, they rarely focus on QoE assessment or on addressing band-limited long-delayed networks.

In this work, we propose congestion avoidance over RTP in conjunction with video adaptation, specifically for QoE in satellite networks. We use utility-based optimization approach, based on [8, 9]. Related work, as in [10, 11], is driven by QoS performance objectives. Other approaches use parameterizations from standardized quality of video to obtain video quality adaptive algorithms [12]. Finally, mappings of subjective QoE metrics are also used as optimization functions [13]. The drawback of the aforementioned approaches is the heterogeneity in the choices for mapping and scenarios, which may not be reproducible or generalized for broader adoption. Furthermore, the existing approaches do not address long roundtrip times, which render them unsuitable for our scenario.

Network Coding for Improving Reliability. In general, wireless systems (specifically satellite systems) suffer from packet erasures. These erasures are due to poor wireless reception conditions and channel fading among others, which the adaptive coding schemes at the physical layer cannot cope with. State-of-the-art video codecs include error concealment features for robustness against erasures; however, they only suffice for short temporal error propagation and may not handle more severe losses [14].

Existing erasure recovery strategies primarily include retransmissions or the use of redundant packets. Retransmission based schemes (e.g., ARQ and TCP) may provide perfect erasure recovery. However, the increase in delay and overhead due to per packet feedback can decrease throughput, especially when round trip times are high. Rateless coding schemes like Luby transform (LT) codes [15] and Raptor codes [16] can generate a fountain of redundant packets and are especially popular for reliable transmission of large files. However, these codes are not efficient with small block sizes, which is the usual structure in video streams [16].

In this work, we examine the use of block coding with random network coding (RNC) to improve reliability. RNC [17, 18] allows mixing of packets to send a fixed amount of redundant packets such that there is sufficient protection guaranteed and the source packets are recovered without the need of feedback. It also provides the inherent possibility of reencoding at intermediate nodes (which is missing in the traditional block coding schemes). The use of RNC for reliable communications in wireless networks was first studied in [19] where RNC was shown to achieve maximum throughput for both unicast and multicast communication with packet erasures. In addition, [2022] discuss the use of RNC for reliability.

In this paper, we focus on systematic random linear network coding (SRNC). SRNC provides an erasure recovery performance similar to maximum distance separable (MDS) codes such as Reed Solomon (RS) codes [23]. In addition, with systematic codes input data is embedded in the encoded output, thereby reducing decoding overhead at the receiver side. Furthermore, the inherent random structure of SRNC makes progressive decoding possible, which improves packet recovery time as compared to using RS codes. This is an advantage in long-delayed scenarios.

QoE and Semantics. QoE is a multidisciplinary field that aims to understand the degree of human satisfaction with an application or service. General QoE models for telecommunications integrate different aspects into a holistic view of QoE [24]. A thorough review of general purpose QoE models and QoE management for wireless networks can be found in [25].

With respect to QoE for video in particular, several features have been studied to improve user’s experience in streaming, such as video coding parameters [26] or temporal impairments [27, 28]. Such solutions focus separately on erasure protection solutions for lossy networks or on dynamic rate adaptation for best effort cases [13, 29].

In contrast to the aforementioned approaches, we base the notions of QoE upon [30], focusing on (1) system influential factors on QoE that relate to the networking scenario at hand and (2) perceptual features in QoE for video that will guarantee delivery of valuable information. Moreover, we follow QoE versus QoS correlation modeling approaches [31], where our aim is high temporal (related to congestion avoidance) and spatial (related to reliable transmission) QoE procurement.

Finally, we propose an additional dimension to our framework that targets specific user demands for situational awareness. In multimedia, “classic” semantics deals with heterogeneous metadata that sensors observe and/or tag when capturing video. It has applications in information retrieval, integration, and aggregation of varied data types such as semantic-aware delivery of multimedia [32]. Furthermore, semantic tagging describing pure observations is used in computer-based systems with artificial intelligence to perceive and abstract situations [33]. Rather than doing perception through classic semantics, we propose a novel human-analysis-driven perceptual semantics approach to tag videos based on the spatial/temporal characteristics of the video a user is perceiving. This provides a mechanism to specifically target and improve the user’s perceptual needs and enhance situational awareness.

Contributions. The main contributions of this paper can be summarized in terms of the following novel aspects:(i)Novel scenario: we concentrate on the use of video to fulfill beyond recreational needs, for example, situational awareness in critical situations; hence the key issue is the iterative use of video over scarce bandwidth. Under these circumstances, video transmission cannot be thought of as a standard streaming solution for domestic use over the internet, nor can it rely on well-known and available encoders or solutions, but on very robust and well-controlled adaptation and coding.(ii)Novel decoupling of time/space for the video adaptation/coding: we address the user’s specific perceptual demands and map, in time and in space, the corresponding network triggers that degrade the user’s perceptual awareness. Based on this mapping, we propose a robust and controlled optimization by decoupling the time and space domains. In addition, this approach proves to be useful in tackling systematically the stringent restrictions of our communications scenario and meets the user’s demands.(iii)Novel perceptual semantic level: we propose a novel perceptual semantics dimension that is intrinsically related to the situational awareness scenario and the end-user driven nature of our approach. Such problems have not been addressed by current state-of-the-art video streaming solutions that focus mainly on communication for recreational needs. The end-user is involved in interactive adaptation through perceptual semantics feedback such that specific user-demanded perceptual spatiotemporal enhancements are possible. Furthermore, it is compliant to current content-aware networking trends.(iv)The use of network coding: network coding is used to improve reliability. It proves to perform similar to MDS codes, with several advantages over MDS codes including reduced decoding complexity, smaller delay, and flexibility to perform adaptive coding. In addition, this scheme allows the extension of coding at intermediate nodes in more complex networking scenarios, providing better performance in terms of throughput and reliability.(v)Experimental validation: we show through a time/space graphical analysis that the joint optimizations achieve planar, homogeneous performance with high values of QoE metrics. This performance is achieved regardless of both erasures and congestion degrading the network. In addition, both optimizations guarantee good performance of the perceptual semantics level to meet the user’s perceptual demands for situational awareness. Moreover, our framework has proven to be of high relevance in realistic scenarios [1].The rest of the paper is organized as follows. In Section 2 we present the scenario. In Section 3, we discuss the system model. In Section 4, we present the QoE optimization in the time domain. Section 5 discusses the QoE optimization in the space domain. Section 6 presents the integration of perceptual semantics into the framework. In Section 7, we present our experimental results. Finally, we present concluding remarks in Section 8.

Notation. Let be a finite field. We denote as the set of all matrices with entries in and as the set of all column vectors with entries in . Boldface uppercase letters are used to denote matrices and boldface letters to denote column vectors. is used to denote identity matrix. The notation represents the set that contains distinct columns of identity matrix . denotes the gradient with respect to .

2. Scenario

We consider point-to-point live streaming of user generated video content for beyond recreational purposes in challenging communications scenarios. The end-user is receiving the live stream and has the possibility of demanding enhancements of video features interactively. The received stream is helping the user improve his/her awareness of the situations depicted in the video, with no use of artificial intelligence in the perceptual and awareness processes [34, 35]. Emergencies, monitoring, or telemedicine are an example of potential scenarios.

2.1. Spatiotemporal Abstraction of Video Services Beyond Recreational

We propose spatiotemporal abstraction that is closer to the perceptual demands of the user. This abstraction is inspired by situational awareness scenarios [36] and diverges from traditional spatiotemporal concepts in video coding, for instance.

The spatial abstraction refers to precise time-space accounts of an ongoing situation such as precision of details and accuracy for identification in a crowd. The temporal abstraction refers to insights in the temporal aspects of dynamically changing situations such as evolution of events and temporal tracking [34].

2.2. Networks in Emergency Scenarios

We assume a sender with access to a band-limited communications network. We consider portable/mobile IP-based satellite services, such as the broadband global area network (BGAN) [37], often used in emergency scenarios, provided by a network of geostationary satellites. These types of services, while ubiquitous, offer limited broadband capacity compared to state-of-the-art wireless terrestrial mobile networks. In addition, inherent long propagation delays are present (in particular in geostationary satellite topologies) and losses from the wireless medium render the network unreliable. As a generalized case we consider best effort provisioning since guaranteed services may not be available. Congestion is thus present.

3. System Model

3.1. QoS/QoE Modeling by Time/Space Decoupling
3.1.1. System Influential Factors Indicators

Definition 1. Quality of Service, QoS, is the ability of the network or service to provide or guarantee a certain level of performance for a data flow.
We consider the following QoS metrics to quantify the influence of the effects of congestion and erasures in the best effort satellite scenario.

Definition 2. Erasure rate is a random variable that follows an i.i.d random process. It represents packet erasure rate due to channel fading in wireless links.

Definition 3. Congestion-induced erasure rate is a random variable that follows an i.i.d random process. It represents packet erasure rate due to congestion in best effort wireless networks.

Definition 4. Network delay is used as an indicator of congestion.

Definition 5. Degree of congestion represents how congested the network is with respect to the maximum available rate offered .  , where and is the available rate to the user at any given time. A value of tending to 0 indicates severe congestion, while indicates no congestion. ( depends on the underlying network (i.e., for the BGAN network in the best effort mode  kbps)).

3.1.2. QoE Framework

We first present the framework used to decompose the system and perceptual aspects of the scenario in Section 2, according to standard QoE definitions.

Definition 6. Quality of experience, QoE, is the degree of delight or annoyance of the user of an application or service [30]. Continuing with the taxonomy proposed by [30], QoE is decomposed into influential factors and perceptual features.

(a) QoE System Influential Factors. These factors signify the technical aspects affecting quality of the application or service, such as media capture, coding, transmission, and playback. Such factors may lead to noticeable degradations such as artifacts, blockiness, and freezes. In the scenario considered in this paper, the QoE system influential factors are linked to the underlying network performance, for example, the best effort wireless satellite network and its QoS.

Other influential factors affecting QoE are context and human factors. Human factors, surrounding emotional and sociological backgrounds, are out of scope of this paper. Context aspects can be a natural extension of the work we present here.

(b) Perceptual Features of QoE. These features are the perceivable characteristics of a user’s experience contributing to the overall quality [30]. These features are directly linked to our spatiotemporal abstraction for video services presented in Section 2. Henceforth, we distinguish a time and space domain decomposition based on the spatiotemporal perceptual features in QoE.

In the space domain, we refer to user’s dissatisfaction due to a lack in accuracy, artifacts in the video caused by packet erasures, and source coding distortion, among others. In the time domain, we refer to user’s dissatisfaction due to persistent freezes in the video playback that prevent tracking dynamically changing situations. In Section 7.2 we discuss the metrics used to measure the spatiotemporal perceptual features of QoE.

3.1.3. QoS/QoE Mapping

Figure 1 summarizes the proposed time-space decomposition. Our analysis is as follows.

Figure 1: Scenario-specific QoE framework with decoupling in time and space domains of QoE.

Congestion affects QoE primarily in the time domain, inducing freezes in video playback. If congestion can be tracked at the transport layer, rate adaptation to the network’s available rate can be performed and QoE in the time domain will be improved.

Erasures affect QoE in the space domain, inducing artifacts in video. Channel coding in the network layer can help recover from erasures, thereby improving QoE in the space domain.

By mapping congestion to the time domain and erasures to the space domain, we are able to propose decoupled solution for the joint problem affecting our scenario. Hence, we propose two QoE-driven optimizations, jointly operative but working separately, one for the time domain and the other for the space domain, to work at the transport and network layers, respectively.

A decoupled solution provides advantages in terms of flexibility of the design since the formulation and performance evaluation of the two optimizations can be treated separately. A potential concern is whether the optimizations can affect one another when they are operating at the same time. In Section 7 we show that, under reasonable assumptions, this cross-influence is minimal.

3.2. Perceptual Semantics Model

Classic semantics based approaches typically use unprocessed sensorial observations [33]. Our proposed perceptual semantics approach represents more complex abstractions of a viewed scene. Based on the spatial/temporal abstraction for video services in Section 2 and its mapping to perceptual features in QoE as shown in Figure 1, our proposal is to utilize the end-user’s (analyst) perception, to do semantic tagging that enables an enhancement of the received video stream signal tailored to the user’s demand.

We propose perceptual tagging that indicates the spatial/temporal predominance according to the level of perception of the user. A tag indicating predominance of temporal features implies that the user is perceiving a situation that demands more attention to the dynamics of the scene (e.g., rapid movements). On the other hand, predominance of spatial features indicates moments of less movement but densely overloaded frames, which requires more detail to identify features.

In scenarios where perception is not achieved by artificial intelligence, human analysis interprets the sensory information (i.e., perceiving). Hence, we propose semantic tagging to be performed by the user, as he/she is ultimately the one perceiving and foreseeing what might be of interest in the video.

3.3. Topology

We consider a point-to-point scenario where the underlying satellite network topology can have several intermediate nodes. In this paper we consider channel coding for reliability only at the source node. However, our system model can be extended to allow network coding at intermediate nodes of the network (this is out of the scope of this paper and a part of ongoing work). Figure 2 shows the overall block diagram of the source-destination topology and our proposed solution.

Figure 2: Scenario.
3.4. Cross-Layer Optimization

As seen in Figure 1, we propose a decoupled cross-layer QoE-driven optimization framework consisting of two types of optimization, one in the time domain and the other in the space domain of QoE.

In the time domain, as shown in Figure 3(a), we use an online adaptation strategy that uses end-to-end feedback at transport layer to cope with congestion. Network delay and congestion-induced erasures can be inferred from the feedback and used to estimate . The application layer rate to be used by the video streaming application is . As a result, the transmission rate at the network layer is (we consider overhead due to layer encapsulation to be negligible when calculating the rates) after optimization, which matches the application layer rate.

Figure 3: Proposed cross-layer optimization framework.

Figure 3(b) shows the integration of the QoE optimizations in the time and space domains. The available rate is first estimated online by the optimization in the time domain. is used as input to the optimization in the space domain to obtain the optimal code rate for erasure protection using RNC coding. As a result, the application layer rate is adapted to and the transport layer packets are encoded using RNC at a sublayer of the network layer. Finally, the IP sublayer transmits at a rate . The optimization in the space domain can be performed offline, and look-up tables can be available online with optimal values for a certain set of input values.

The online adaptation strategy, resulting in dynamic rate adaptation, fulfills two purposes, namely, (1) congestion control at transport layer and (2) online adaptation of the video source for maximized QoE.

Note that RNC is chosen to be at the network layer in order to enable the possibility of coding at intermediate nodes in future work. Network layer packets are accessible at intermediate nodes; hence coding at this layer would be more efficient in our model.

3.5. Matricial System Model

We consider the frame structure of standard state-of-the-art video codecs to model the source. Coded frames are grouped into groups of pictures (GoPs). Each GoP has three types of frames, namely I, P, and B frames, each of different importance. Network coding can be used to provide unequal protection (UEP) for these different frames [9]; however, in this work, we consider equal protection and focus on the allocation of redundant packets to the complete GoP block.

We consider each GoP to have a fixed number of frames . The frame rate is such that the codec outputs each GoP in a fixed time . We denote as the total number of GoP’s output by the codec during the entire streaming session such that .

The codec outputs the th GoP, coded at application layer rate , for at time . Although is fixed, frame sizes vary depending on the . For the th GoP, each frame is fragmented into multiple packets of equal length (in bits) and delivered from the transport layer to the () layer for end-to-end delivery. We denote as the total number of packets from the th GoP. We drop the index for simplicity in formulation; however, as the source is time varying, coding parameters depending on can also vary from one GoP to another.

We define as containing all the packets from a GoP. Each packet is a column vector of symbols where is a function of field size with packet length , given by , and is a prime number called the characteristic of the field.

Encoding is done at the RNC layer as shown in Figure 3(b). The encoding process is linear such that the coded packets corresponding to source packets are given bywhere . is the corresponding generator matrix of linear code, with , where is the code rate used for encoding.

Specifically in SRNC, the coded packets are composed of the embedded source packets (systematic phase) and the redundant packets, product of linear combination of the source packets (nonsystematic phase). Hence, the generator matrix results in , where is a matrix with random coefficients from the finite field . The linear combining is random and it is not constrained to specific combination of coding parameters. This allows us to have flexibility in choosing coding parameters which may vary from one GoP to another.

After the addition of IP headers, these IP packets are transmitted to the destination over an erasure channel where the packets can be erased. We denote the channel function , which maps encoded packets to received packets. We denote the received unit by the matrix such that each received packet is a column vector of symbols. In our case, the channel model is linear and we have , with .

The matrix represents the erasure of packets, consisting of all the columns of except the columns if the th column/packet is erased by the channel. The channel matrix deletes the packets of which are lost and hence consists only of received packets.

4. Optimization in the Time Domain

In this section we present the formulation and implementation of the QoE-driven optimization in the time domain, as part of the cross-layer model in Figure 3. In our QoE decoupling approach, we have identified congestion with freezes in the video playback and hence propose an optimization for improved QoE in the time domain. The implementation of this optimization results in a dynamic rate adaptation algorithm.

4.1. Formulation

Consider a best effort wireless scenario with a network varying over time .

The general formulation of our objective optimization problem is presented in (2), where the utility function is dependent on the QoS parameters. The QoS parameters considered are the transmission rate , delay , and congestion-induced erasures , all of them varying with time (, and are function of time ; for clarity in stating the optimization problem we have dropped in (2)–(5)).   is the upper bound on maximum rate offered by the network. The available rate offered by the network may vary over time and is unknown to the user. Consider

Consider an additive model, where the utility is composed of two functions, namely, one representing QoE’s improvement with increasing assignment of network resources and a second one representing the dynamics degrading the network in the best effort scenario. Considerwhere is a concave function, defined in (4) based on the logarithmic mappings from QoS to QoE. Studies have shown that if the rate is increased in a controllable fashion (e.g., by increasing the application layer rate of the video), QoE behaves with a logarithmic relationship [39]. Consider, on the other hand, expresses the penalizing effect of a congested network scenario, where injecting higher rate than the currently available one for the user () translates into accumulating delay and eventually an overflow of network buffers leading to packet losses. Hence, we formulate as a bilinear function of and inNotice that we define the function to strengthen or weaken the effect of in the overall optimization depending on the level of congestion perceived, as proposed in [40, 41], for flow control applications.

4.2. Implementation as Dynamic Rate Adaptation
4.2.1. Solution to the Optimization Problem

Proposition 7. The optimization problem stated in (2) where the utility function is defined as in (3) is solved using the discrete rate update algorithm (6), to find the value of at time , for , where is the network sampling time, is the step size, and is the gradient with respect to . Consider

Proof. First we prove that is concave and hence an optimal value that solves (2) exists. The function from (4) is strictly concave increasing with , while is concave, decreasing with . The sum of concave functions is concave; hence is concave and an optimal that solves (2) exists. Further, the gradient ascent method can be used to find the optimal , where is varying over time in the direction of the positive gradient of . In practice, rate updates happen in discrete time, and if we consider sampling time , the rate control update is expressed as in (6).

Observe that changes with the current network conditions at time ; as a result, knowledge of QoS levels at the transport layer is needed to solve the optimization problem. Such knowledge of the network is based on feedback from the receiver end. If we consider feedback delay, the measurements represent QoS levels at delayed points. This is especially true in the case of long delay networks, such as satellite networks, where propagation delay is noticeable.

Proposition 8. In the case of a delayed network, with propagation delay , the algorithm in (6) that solves (2), with defined from (3), (4), and (5), is expressed as in

Proof. The algorithm is triggered when new network measurements are available at the sender side; however, those are measurements corresponding to the network state at time . Furthermore, if we consider the rate control update with sampling time to be greater than the network’s roundtrip time (), we can assume that the receiver is able to report on network changes related to the last rate control action from the sender at time . Consequently, we can express (6) as , where the gradient of is evaluated at time and the gradient of is evaluated at time . Substituting and for (5) and (4) we obtain (7).

The function is chosen such that the response of the adaptation depends on the current measurements indicating congestion and can react faster to increasing delay constraints and packet loss as described by (8), where . Note that responds to increases in , which are accompanied with increases in delay . Hence other sources of packet erasures not related to congestion will not trigger a change in the rate control update. Consider

The delay-driven rate update obtained from (6) using the value of according to (8) provides a smooth (an advantage to user’s QoE [29]) output that is also capable of reacting fast to severe degradations in QoS. and are chosen for a desired response time, while and correspond to upper bound limits to and , set according to application requirements.

4.2.2. Cross-Layer Aspects and Practical Issues

The optimization proposed in this section is used within the whole cross-layer framework in Figures 3(b) and 3(a), such that . Further, in order to maintain coherence of our model in time and avoid synchronization issues at different layers, we assume that . The obtained application layer rate for the th GoP is thus invariant for the duration of the whole GoP.

Cross-layer feedback from the receiver is provided by the real-time control protocol, RTCP, (RFC 3550) with a frequency of reporting of . In order to obtain and to estimate , the following fields from both sender and receiver RTCP reports are required (following RFC 3550 standard): fraction lost, delay since last report (DLSR), and last sent report (LSR).

The obtained algorithm results in high granularity rate adaptation. Therefore it requires a video codec capable of performing on-the-fly encoding with fine granularity. The standard codec H.264/AVC offers such features, with possibility of adaptation of its quantization parameters (QP) at encoding time. The VP8 codec also offers such capabilities [42], with the option of real-time encoding with on-the-fly reconfiguration of application layer rate. The extension of the H.264 codec for scalable video coding (SVC) could also be considered under certain assumptions. SVC would offer more coarse granularity in achieving the rate , depending on the combinations of temporal/spatial/amplitude scalability layers. Therefore, additional buffering might be needed in order to diminish potential impact on congestion. Further, computational complexity during real-time coding of all layers would increase.

5. Optimization in the Space Domain

In our decoupling approach, we identify erasures with artifacts in the video to be solved using an optimization in the space domain. In this section, we present the formulation and implementation of such QoE-driven optimization, as part of the cross-layer model in Figure 3.

The objective of the optimization in the space domain is to optimize application rate and code rate , in order to use SRNC to cope with erasures with maximized QoE of video.

5.1. Formulation

Let us consider to be the available rate estimated using the algorithm in (6). In order to protect the video stream from network erasures, SNRC coding will be used with a certain allocated code rate . A low value of implies more erasure protection, at the expense of a lower rate for the application layer . Given that a lower results from higher compression rates, QoE in the space domain is damaged with low values of .

Hence, we propose in (9) maximizing QoE by maximizing , such that SRNC is used with an optimal code rate that guarantees a residual erasure rate . Considerwhere is the residual erasure rate of SRNC with field size .

We target an offline solution to (9) in order to obtain the optimal values ( and ) corresponding to all the possible estimated available rates . A look-up table with these values is generated. As is time varying and is estimated from feedback, the look-up table is accessed online and optimal and are obtained corresponding to . Consider

5.2. Implementation Based on SRNC

In this section, we present the implementation of SRNC and its performance.

Following the matricial model in Section 3, we have and depending on and . Hence the residual erasure rate is expressed as a function of and . Moreover, as noted earlier, we receive consisting of coded packets for each GoP. From the received unit , we can obtain the source unit , when (i) is known at the receiver and (ii) . If these two conditions are fulfilled, then at the destination source packets are obtained by .

In order to recover we use pseudorandom network coding [43]. The index of the codebook is sent along with each coded packet (the columns of ). The receiver, which has the same codebook, can recover with this index. The advantage of this option is low overhead (only the index per packet) compared with the overhead of sending the coefficients (each column of ) attached to each coded packet [17].

Gauss-Jordan elimination (instead of Gaussian elimination) is used for progressive decoding. As a result, it is not necessary to wait for the complete block and packet recovery time is reduced. In comparison, state-of-the-art RS codes require a complete block to start decoding.

The source packets are finally recovered if . Let us denote as the probability of successfully decoding the original packets. Based on , we evaluate the residual erasure rate for SRNC in the next proposition.

Proposition 9. If SRNC is used with field size and source packets (from the systematic phase) with redundant packets (from the nonsystematic phase) are transmitted over the erasure channel with random erasure rate, , then the residual erasure rate for SRNC is given bywhere is given by (10).

Proof. We will first evaluate . Let and be the number of packets received from the systematic and nonsystematic phases, respectively. , , and the dimensions of are . The probability of receiving all the source packets from the systematic phase, , is . Once all the source packets are received, redundant packets are not needed. If only some of the original packets are received, , then at least packets are required from the nonsystematic phase for successful decoding. The probability of matrix having full rank when columns are independent is given by using the model in [44]. Combining both cases, and , we have given in (10). Once we have the probability of successful decoding of all the packets, the residual erasure rate is simply given by (11).

In Figure 4, we present numerical results illustrating the effect of field size on . For our results, we choose  kbps, bytes, and second (, , and are chosen corresponding to the realistic scenarios which we are considering for experiments in Section 7) such that . We vary the code rate from 0.5 to 1 and consider erasure rate . Our results show that (i) a field size of is enough for SRNC to guarantee performance similar to that of MDS codes and (ii) there is up to 47% gain in code rate, achievable with as compared to for a target erasure rate of . A higher code rate will result in greater budget allocation to the application rate which improves the QoE in the space domain.

Figure 4: Effect of the field size on residual erasure probability.

6. Integration of Perceptual Semantics

We propose the use perceptual semantics to enhance specific perceptual features, following the model introduced in Section 3.2. Further, we integrate perceptual semantics with the joint time and space optimizations.

6.1. Formulation

We focus on using our proposed perceptual semantics for enhancement at source coding level. In single or scalable layer state-of-the-art video encoding, there are three types of resolution:defined-temporal (frame rate), amplitude (quantization step), and spatial (frame size).

We map enhancement of temporal features to higher frame rates and predominance of spatial features to higher spatial and amplitude frame resolution. In this way, dynamics of a scene can be more closely followed (temporal preference) and details of a scene can be better identified (spatial preference). The mapping is intuitive and nonintrusive and relies on the intrinsic architecture of video codecs currently in use to facilitate the video communications.

We propose mapping perceptual semantics to a system quantified with the variable . and express full preference of the spatial and temporal perceptual features, respectively. Intermediate values of represent weighed spatiotemporal preferences.

We denote the feasible set of finite values of frame rate, as , while is the feasible set for the spatial factors. Both are a function of the application layer rate . Note that higher frame rates and frame sizes are possible to attain with higher [45]; hence, the feasible sets and corresponding to higher values of will contain a greater number of possible values that can be chosen from. For example, in the case of scalable video coding, if temporal dyadic scalability is performed, the available values of frame rate contained in would be the base layer frame rate and the frame rates from the enhancement layers. The combination of all layers adds up to the full frame rate, that is, a full 30 Hz frame rate if is sufficient with .

In order to choose the appropriate value of frame rate and resolution according to our mapping of perceptual semantics, we formulate the following optimization function:where and are the normalized values of frame rate and spatial/amplitude resolution with respect to maximum available values set for the application.

Note that the optimization in (12) can be applied to single or scalable layer video coding.

6.2. Implementation

The implementation into the cross-layer optimization model from Section 3.4 is as follows.

The video streaming application uses a state-of-the-art codec such that the frame rate, frame size, and codec rate can be configured on-the-fly. In order to facilitate the role of perceptual semantics, we use a return path to send the tags chosen by the user. A semantics-aware adaptation block at the sender interprets the semantic tags coming from the end-user by mapping it to the proper decisions in (12) and forwarding to the video codec.

We propose the use of semantic web protocols to enable the semantic feedback to the transmitter through the APP-to-APP cross talk of the semantic tagging [46]. At the transport layer, the application-specific information can be encapsulated into RTCP feedback packets compliant with the extended reports defined in RFC4585. This way, the perceptual semantics feedback loop is coherent with the cross-layer optimization.

Our framework complies with the notion of a semantic information-based network. Hence, it is coherent with the content-aware trends in networking where the focus is on the network as a platform for information dissemination rather than simply an enabler of communication links. This framework can be mapped to information-centric networking (ICN) architectures such as publish/subscribe for live video as in [47]. A feasible topology mapping of ICN to our scenario is discussed in [48].

7. Experimental Results

7.1. Experimental Setup

The setup consists of a point-to-point streaming connection. The receiver and sender applications are connected through an emulated network using the NetEM emulator.

7.1.1. Setup

Following Figure 3(b), we describe each block.

At application layer we use the state-of-the-art video codec VP8 [42]. At transport layer, we use the RTP/UDP protocol and a standard implementation of RTCP protocol for feedback. At network layer, each transport layer packet is encapsulated into an IP packet.

The online optimization has been implemented to output a rate control update of with every new RTCP report, according to (7). The offline optimization uses a look-up table to output the optimal and code rate values from the budget rate .

We simulate SRNC coding by adding, for each GoP coming from the transport layer at rate , redundant (dummy) packets such that .

7.1.2. Network Emulation

With respect to erasures, packets are erased at the random rate when no erasure protection is performed. When SRNC is used, packets are erased corresponding to the residual erasure probability of SRNC .

Congestion events are emulated as a drop (step-like) from maximum available rate , which occurs halfway through one streaming session, at . In practice, we use traffic shaping in the NetEm emulator to create the drops in , such that .

7.1.3. Perceptual Semantics

The evaluation of the perceptual semantics approach of Section 6, integrated into the cross-layer optimization model, is performed using a simulation platform where the video streaming application is simulated by generating packets of size encoded at a rate . Network simulation follows the same guidelines as used with the time and space optimizations. All parameters in Table 2 apply, except for those related to the application layer.

We model a user’s semantic tagging from temporal/spatial features with the parameter . may vary over time throughout one single streaming session, such that the sender is receiving feedback of these changes and adapts to them using (12). We assume that these tags are changed by the user every 10 s. We consider time variation of semantic tagging as a user alternating between spatial and temporal tags, each lasting 10 seconds.

Table 1 summarizes the feasible sets for values of frame rate dependent on , in order to solve the algorithm in (12). The values chosen correspond to typical feasible combinations in current state-of-the-art codecs.

Table 1: Feasible sets considered for simulation of perceptual semantics.
Table 2: Parameters in experimental setup for time and space optimizations.
7.1.4. Experiments

Table 2 summarizes the values of the parameters used for the experiments.

Experiments with and without space and time domain QoE optimizations are considered. Each experiment consists of one streaming session lasting seconds. A large value of (3 minutes) helps guarantee statistical significance with respect to erasure rates as well as spatiotemporal variations in the video.

For each experiment, a looped standard video sequence served as the input source. Furthermore, each experiment utilizes a specific value of and . The ranges of values for are 0%–15%, while, for , the range is from 100% to 50%.

The range of values considered for and corresponds to realistic values for an application using a mobile satellite service, such as the BGAN network. Such network offers roughly maximum  kbps in a best effort configuration. The propagation delay corresponding to a GEO-stationary satellite network is also configured in NetEm. The value of was chosen according to 3GPP (3rd Generation Partnership Project) specifications for real-time scenarios.

7.2. Performance Metrics
7.2.1. Spatiotemporal Perceptual Features of QoE

We measure the spatial and temporal perceptual features of QoE, coherent with the framework described in Section 3.1.2 and our spatiotemporal abstraction of the video.

Application layer information, both at the media and the bitstream levels, is collected at the receiver and the sender for offline performance assessment. A frame concealment strategy is used to avoid misalignment of sent/received video and the associated impact on full reference (FR) spatiotemporal video assessment. This implies that a lost frame is replaced in the sequence by the last frame received, using bitstream level data from actual frames sent and received. In addition, bitstream level data also provides frame play-out timestamps.

measures the spatiotemporal perceptual features of video. It is an FR media level metric, considered in [49], where it was shown to exhibit good correlation to subjective metrics. It is defined as , with variables , , , and as follows. is, in our case, the vector with frame-by-frame full reference video quality metric SSIM from each experiment [50]. indicates the mean value function. is the standard deviation and is a weight value. This metric considers the variability of quality throughout the streaming session; hence it is able to represent the impact of time variations in the network.

measures the temporal perceptual features of video. It is a nonreference metric at bitstream level, representing video flow continuity. is defined as the probability that no freezes appear in the video playback. Freezes are defined as events in which the time elapsing between two consecutive frames displayed during video playback exceeds a tolerated threshold . Hence we can define as .

7.2.2. Perceptual Semantics

We define the combined metric to measure tradeoffs of using perceptual semantics with and without cross-layer optimization. It is defined as follows:with and . , where is the average packet loss rate at the receiver. evaluates the performance of the perceptual semantics algorithm to determine whether the algorithm is achieving the user-demanded . The best performance, that is, , occurs when no losses degrade the video (), freezes in playback are minimal (), and the perceptual semantic adaptation matches the one requested by the user ().

7.3. Joint Optimizations in the Time and Space Domains

The purpose of this experiment is to evaluate the performance of the joint optimizations in the time and space domains according to the proposed model in Figure 3(b). Hence, we consider degradations due to both congestions and erasures. Congestion events and erasures are emulated as in the previous sections with the parameters of Table 2 for Experiment (3). For each experiment a different value of and was considered.

We compare the results of the joint optimization with a solution unaware of network dynamics, where the application layer is blind to the network dynamics, the transport layer is not performing any congestion control, and there is no protection against erasures.

7.3.1. Effect on

The three-dimensional QoE plots in Figure 5 show that, for all cases of congestion and erasures tested, the values of the flow continuity metric are all above 0.9 when using both optimizations. This implies that more than 90% of the time, the user is not experiencing freezes during video playback. The complete framework compared to a non-QoE optimized approach has gains of up to 60% in flow continuity.

Figure 5: Flow continuity, for joint QoE in space and time domain.
7.3.2. Effect on

The complete solution achieves in overall a planar surface in , as shown in Figure 6. This implies that, regardless of both erasures and congestion affecting QoE, the combination of the online and offline strategies is able to deliver smooth performance.

Figure 6: metrics for joint QoE in space and time domain.

Moreover, for all cases of congestion and erasures tested, the values of metric are all above 0.9 when using the complete QoE framework, guaranteeing very small variations in quality over time. The gain with respect to a nonoptimized approach is of up to 80%. The cases with higher improvement correspond to higher erasure rates and greater degree of congestion . The gradient of the gains in for higher values of is only dependent on the increase in erasure rates, while, for lower values of , the gain increases jointly as and increase.

These results represent a high QoE in the space domain together with smooth QoE performance throughout the entire streaming session, a characteristic highly valued by end-users. This behavior was observed with all video sequences tested.

7.4. Trade-Offs in the Decoupling Approach

We comment on the trade-offs by analyzing the isolated performance of both time and space optimizations and their effects on the metrics.

7.4.1. Optimization in Time Domain

In this case we consider degradations due to congestion only and compare the performance to a solution that is unaware of such degradations. The results are summarized in Figure 7 for videos coastguard, pedestrian, and foreman. The parameters for this experiment correspond to Experiment (1) from Table 2.

Figure 7: Evaluation of only QoE optimization in the time domain.

(a) Effect on . As can be observed from Figure 7(a), the flow continuity measured with is improved up to 50%. The highest advantage is achieved for lower values of .

(b) Effect on .The online optimization is able to avoid congestion events; hence packet loss due to congestion is minimized, proving that the space-time decoupling premise is valid. As a consequence, the improvements in QoE are not only in time domain metrics but also in the space domain.

It can be observed from Figure 7(b) that the most significant improvements occur for congestion events with in , with improvement of over 100%. For higher values of , the metric also shows a gain from 4.5% to 50% using the QoE optimization in the time domain.

7.4.2. Optimization in the Space Domain

In this experiment we compare the optimization in the space domain (where for each an optimal is obtained to configure and use SRNC) to a nonoptimized strategy with no erasure protection (). This case considers only degradations in the network due to erasures. We assume is constant throughout the entire streaming session (; there is no congestion). We assume the transmission rate , and . For each experiment, there is a corresponding pair of values . The parameters are set as in Table 2 for Experiment (2).

(a) Effect on . By optimizing the rate budget when using SRNC, we ensure that the redundancy added will not congest the network. QoE in the time domain is therefore not affected by the use of SRNC, as is intended in our decoupling approach.

Notwithstanding, SRNC, similar to other block erasure codes, adds delay at encoding/decoding. The systematic characteristic of SRNC as well as the possibility of performing progressive RNC decoding significantly reduces the delays imposed by erasure protection. Therefore, we can assume a reduced start-up delay in the video playback. This small price to pay has a duration not longer than a GoP, , thereby guaranteeing minimal impact of SRNC on QoE in the time domain.

(b) Effect on . Figure 8 shows the results for videos foreman and coastguard, in our three-dimensional analysis of QoE, where we plot QoE metrics versus versus . Using the optimization in the space domain, the main advantage is achieved in scenarios with high available rate and high erasure rates, with up to 38% improvement in metric compared to a nonoptimized strategy. The higher advantage occurs for higher values of .

Figure 8: QoE optimization in the space domain. metric.

We show minimal effects of the decoupling approach on with the following example. Consider the surface representing the case where space optimization is not utilized in Figure 8. In an erasureless scenario (, ), reduction of 20% in has degradation of under 1% in , while reduction in goodput due to erasures represents 40% degradation in . This shows that the spatial perceptual features of QoE are not sacrificed when part of the budget rate is used for erasure protection. This is confirmed by the smooth performance of our solution in Figure 8. Due to the joint operating optimizations in both the time and space domains, we gain benefits in both. In time a rate that avoids congestion is ensured. In space we optimally assign the resources for SRNC () and application layer () such that we obtain a target residual erasure rate and the spatial perceptual features of QoE are preserved.

7.5. Perceptual Semantics with and without Time and Space Optimizations

To our knowledge, there is no similar framework in the literature to match our proposed perceptual semantics framework and hence comparison to solutions that do not have a similar goal would be unfair. Therefore, our results focus on not having such a kind of framework. In order to observe the combined effects of the adaptation through perceptual semantics with the cross-layer optimization, we compare the use of cross-layer optimization to cope with the network constraints to a situation where it is not used.

We analyze the effects of time-varying perceptual tagging, representing a realistic case where the user identifies different situations that demand attention towards temporal or spatial features. These variations are represented as alternations of temporal and spatial tagging. Figure 9 shows the performance in terms of the combined metric .

Figure 9: for time-varying semantic tagging .

In addition to achieving the expected demanded through the use of semantic tagging, the performance is above 80% regardless of the degradations of the network, thanks to the cross-layer optimization. The performance is highly degraded due to congestion as well as erasures when no cross-layer optimization is used, with performance dropping to 40%.

Figure 9 confirms the analysis by showing that the cross-layer optimization preserves the perceptual semantics.

8. Conclusions

In this work, we proposed a solution to deliver point-to-point video services in best effort satellite networks for purposes beyond recreational, such as for situational awareness. We used QoE framework to decouple the problems inherent to the scenario, relating congestion with freezes in the time domain and packet erasures with artifacts in the space domain. Both impairments degrade the QoE of video and as a result the ability of video to help gain situational awareness. Our decoupled approach facilitates the design to optimize QoE both in the time and in the space domains, thereby providing a feasible solution for dynamic adaptive streaming tailored to the scenario’s needs. As a consequence of decoupling and tackling these two problems separately, we have performed a time/space graphical analysis with varying network conditions in form of congestion and erasures. Furthermore, driven by the temporal-spatial abstraction of video and its perceptual features, we presented a novel model for perceptual semantics, based upon the user’s demands. We also proposed the framework to be integrated into an interactive video adaptive solution, for user situational awareness. We discussed how to practically implement perceptual semantics into an adaptive loop that works with underlying cross-layer optimization. Our experimental results showed the benefits of this decoupled approach in terms of objective QoE metrics. We were able to achieve homogenous high performance, regardless of both erasures and congestion degrading the network. Our simulation results also showed how perceptual semantic tagging achieved the expected user demands while the underlying cross-layer optimization preserved performance. Future work includes the extension of our analysis to the general network where intermediate nodes can perform coding for higher reliability and throughput. Furthermore, other aspects of QoE, such as context, can be studied within our decoupled QoE framework. In addition, extensions of perceptual semantics in the ICN context will be pursued. The ICN umbrella allows the future consideration of our scheme for live multicast dissemination to a number of users assessing simultaneously an ongoing critical mission. The last mile networking elements in ICN could be in charge of the multicast distribution. Moreover, we will study more pertinent QoE metrics to match user’s satisfaction when using perceptual semantics.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

Part of the work presented in this paper was developed within the project GEO-PICTURES [1], funded under the FP7 framework by the European Commission. The scheme the authors propose in this paper will be extended in the recently awarded project GEO-VISION, funded under the H2020 framework by the European Commission.

References

  1. FP7 Project GEO-PICTURES, http://www.geo-pictures.eu.
  2. B. Wang, J. Kurose, P. Shenoy, and D. Towsley, “Multimedia streaming via TCP: an analytic performance study,” ACM Transactions on Multimedia Computing, Communications and Applications, vol. 4, no. 2, article 16, pp. 1–22, 2008. View at Publisher · View at Google Scholar · View at Scopus
  3. S. Floyd, M. Handly, J. Padhye, and J. Widmer, “TCP friendly rate control (TFRC): protocol specification,” IETF RFC 5348, 2008. View at Google Scholar
  4. H. Seferoglu, A. Markopoulou, U. C. Kozat, M. R. Civanlar, and J. Kempf, “Dynamic FEC algorithms for TFRC flows,” IEEE Transactions on Multimedia, vol. 12, no. 8, pp. 869–885, 2010. View at Publisher · View at Google Scholar · View at Scopus
  5. H.-P. Shiang and M. van der Schaar, “A quality-centric tcp-friendly congestion control for multimedia transmission,” IEEE Transactions on Multimedia, vol. 14, no. 3, pp. 896–909, 2012. View at Publisher · View at Google Scholar · View at Scopus
  6. V. Singh, J. Ott, and I. D. D. Curcio, “Rate-control for conversational video communication in heterogeneous networks,” in Proceedings of the 13th IEEE International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM '12), pp. 1–7, June 2012. View at Publisher · View at Google Scholar · View at Scopus
  7. L. de Cicco, G. Carlucci, and S. Mascolo, “Experimental investigation of the google congestion control for real-time flows,” in Proceedings of the 5th ACM SIGCOMM Workshop on Future Human-Centric Multimedia Networking (FhMN '13), pp. 21–26, ACM, August 2013. View at Publisher · View at Google Scholar · View at Scopus
  8. M. A. Pimentel-Niño, M. A. Vázquez-Castro, and H. Skinnemoen, “Optimized ASMIRA advanced QoE video streaming for mobile satellite communications systems,” in Proceedings of the 30th AIAA International Communications Satellite Systems Conference, September 2012.
  9. M. A. Pimentel-Niño, P. Saxena, and M. A. Vázquez-Castro, “QoE driven adaptive video with overlapping network coding for best effort erasure satellite links,” in Proceedings of the 31st AIAA International Communications Satellite Systems Conference (ICSSC '13), Florence, Italy, October 2013. View at Publisher · View at Google Scholar · View at Scopus
  10. M. Chiang, S. H. Low, A. R. Calderbank, and J. C. Doyle, “Layering as optimization decomposition: a mathematical theory of network architectures,” Proceedings of the IEEE, vol. 95, no. 1, pp. 255–312, 2007. View at Publisher · View at Google Scholar · View at Scopus
  11. D. Pradas and M. A. Vázquez-Castro, “NUM-based fair rate-delay balancing for layered video multicasting over adaptive satellite networks,” IEEE Journal on Selected Areas in Communications, vol. 29, no. 5, pp. 969–978, 2011. View at Publisher · View at Google Scholar · View at Scopus
  12. R. N. Vaz, B. W. M. Kuipers, and M. S. Nunes, “Video quality optimization algorithm for video-telephony over IP networks,” in Proceedings of the IEEE 21st International Symposium on Personal Indoor and Mobile Radio Communications (PIMRC '10), pp. 2545–2550, September 2010. View at Publisher · View at Google Scholar · View at Scopus
  13. O. Habachi, Y. Hu, M. van der Schaar, Y. Hayel, and F. Wu, “MOS-based congestion control for conversational services in wireless environments,” IEEE Journal on Selected Areas in Communications, vol. 30, no. 7, pp. 1225–1236, 2012. View at Publisher · View at Google Scholar · View at Scopus
  14. T. Schierl, M. M. Hannuksela, Y.-K. Wang, and S. Wenger, “System layer integration of high efficiency video coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 22, no. 12, pp. 1871–1884, 2012. View at Publisher · View at Google Scholar · View at Scopus
  15. M. Luby, “LT codes,” in Proceedings of the 43rd Annual IEEE Symposium on Foundations of Computer Science, pp. 271–280, IEEE, Vancouver, Canada, November 2002. View at Publisher · View at Google Scholar
  16. A. Shokrollahi, “Raptor codes,” IEEE Transactions on Information Theory, vol. 52, no. 6, pp. 2551–2567, 2006. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  17. P. A. Chou, Y. Wu, and K. Jain, “Practical network coding,” in Proceedings of the 41st Annual Allerton Conference on Communication, Control and Computing, October 2003.
  18. T. Ho, M. Medard, R. Koetter et al., “A random linear network coding approach to multicast,” IEEE Transactions on Information Theory, vol. 52, no. 10, pp. 4413–4430, 2006. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  19. D. S. Lun, M. Medard, and M. Effros, “On coding for reliable communication over packet networks,” in Proceedings of the 42nd Annual Allerton Conference on Communication, Control, and Computing, September 2004.
  20. H. Balli, X. Yan, and Z. Zhang, “On randomized linear network codes and their error correction capabilities,” IEEE Transactions on Information Theory, vol. 55, no. 7, pp. 3148–3160, 2009. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  21. D. Vukobratovic, C. Khirallah, V. Stankovic, and J. S. Thompson, “Random network coding for multimedia delivery services in LTE/LTE-Advanced,” IEEE Transactions on Multimedia, vol. 16, no. 1, pp. 277–282, 2014. View at Publisher · View at Google Scholar · View at Scopus
  22. P. Saxena and M. A. Vázquez-Castro, “Network coded multicast and multi-unicast over satellite,” in Proceedings of the International Conference on Advances in Satellite and Space Communication (SPACOMM '15), Barcelona, Spain, April 2015.
  23. P. Saxena and M. A. Vázquez-Castro, “Network coding advantage over mds codes for multimedia transmission via erasure satellite channels,” in Proceedings of the International Conference on Personal Satellite Services (PSATS '13), 2013.
  24. K. U. R. Laghari, K. Connelly, and N. Crespi, “Toward total quality of experience: a QoE model in a communication ecosystem,” IEEE Communications Magazine, vol. 50, no. 4, pp. 58–65, 2012. View at Publisher · View at Google Scholar · View at Scopus
  25. S. Baraković and L. Skorin-Kapov, “Survey and challenges of QoE management issues in wireless networks,” Journal of Computer Networks and Communications, vol. 2013, Article ID 165146, 28 pages, 2013. View at Publisher · View at Google Scholar · View at Scopus
  26. S. J. Hussain, R. J. Harris, and G. A. Punchihewa, “Dominant factors in the content domain that influence the QoE of an IPTV service,” in Proceedings of the 1st IEEE TENCON Spring Conference, pp. 572–577, IEEE, April 2013. View at Publisher · View at Google Scholar · View at Scopus
  27. T. Hossfeld, S. Egger, R. Schatz, M. Fiedler, K. Masuch, and C. Lorentzen, “Initial delay vs. interruptions: between the devil and the deep blue sea,” in Proceedings of the 4th International Workshop on Quality of Multimedia Experience (QoMEX '12), pp. 1–6, July 2012. View at Publisher · View at Google Scholar · View at Scopus
  28. A. ParandehGheibi, M. Médard, S. Shakkottai, and A. Ozdaglar, “Avoiding interruptions—QoE trade-offs in block-coded streaming media applications,” in Proceedings of the IEEE International Symposium on Information Theory (ISIT '10), pp. 1778–1782, Austin, Tex, USA, June 2010. View at Publisher · View at Google Scholar · View at Scopus
  29. G. Tian and Y. Liu, “Towards agile and smooth video adaptation in dynamic HTTP streaming,” in Proceedings of the 8th ACM International Conference on Emerging Networking EXperiments and Technologies (CoNEXT '12), pp. 109–120, ACM, December 2012. View at Publisher · View at Google Scholar · View at Scopus
  30. P. L. Callet, S. Moeller, and A. Perkis, “Qualinet white paper on definitions of quality of experience, version 1.1,” Tech. Rep., European Network on Quality of Experience in Multimedia Systems and Services (COST Action IC 1003), Lausanne, Switzerland, 2012. View at Google Scholar
  31. M. Alreshoodi and J. Woods, “Survey on QoE/QoS correlation models for multimedia services,” International Journal of Distributed and Parallel Systems, vol. 4, no. 3, p. 53, 2013. View at Google Scholar
  32. J. Thomas-Kerr, C. Ritz, and I. S. Burnett, “Semantic-aware delivery of multimedia,” in Proceedings of the 9th International Symposium on Communications and Information Technology (ISCIT '09), pp. 1498–1503, September 2009. View at Publisher · View at Google Scholar · View at Scopus
  33. C. Henson, A. Sheth, and K. Thirunarayan, “Semantic perception: converting sensory observations to abstractions,” IEEE Internet Computing, vol. 16, no. 2, pp. 26–34, 2012. View at Publisher · View at Google Scholar · View at Scopus
  34. F. Bergstrand and J. Landgren, “Information sharing using live video in emergency response work,” in Proceeding of the 6th International ISCRAM Conference, Gothenburg, Sweden, May 2009.
  35. S. S. Krupenia, C. Aguero, and K. C. Nieuwenhuis, “The value of different media types to support command and control situation awareness,” in Proceedings of the 9th International ISCRAM Conference, Vancouver, Canada, April 2012.
  36. M. R. Endsley, “Theoretical underpinnings of situation awareness: a critical review,” in Situation Awareness Analysis and Measurement, pp. 3–32, 2000. View at Google Scholar
  37. Inmarsat BGAN, http://www.inmarsat.com/service/bgan/.
  38. 3GPP, “Policy and charging control architecture,” 2010.
  39. T. Hosfeld, M. Fiedler, and T. Zinner, “The QoE provisioning-delivery-hysteresis and its importance for service provisioning in the Future Internet,” in Proceedings of the 7th EURO-NGI Conference on Next Generation Internet (NGI '11), pp. 1–6, IEEE, Kaiserslautern, Germany, June 2011. View at Publisher · View at Google Scholar
  40. Z. Rosberg, “Rate control with end-to-end delay and rate constraints,” in IEEE INFOCOM Workshops, pp. 1–6, April 2008. View at Publisher · View at Google Scholar · View at Scopus
  41. T. Alpcan and T. Başar, “A globallly stable adaptive congestion control scheme for internet-style networks with delay,” IEEE/ACM Transactions on Networking, vol. 13, no. 6, pp. 1261–1274, 2005. View at Publisher · View at Google Scholar · View at Scopus
  42. J. Bankoski, P. Wilkins, and Y. Xu, “Technical overview of VP8, an open source video codec for the web,” in Proceedings of the 12th IEEE International Conference on Multimedia and Expo (ICME '11), Barcelona, Spain, July 2011. View at Publisher · View at Google Scholar · View at Scopus
  43. C.-C. Chao, C.-C. Chou, and H.-Y. Wei, “Pseudo random network coding design for IEEE 802.16m enhanced multicast and broadcast service,” in Proceedings of the IEEE 71st Vehicular Technology Conference (VTC '10), pp. 1–5, 2010.
  44. O. Trullols-Cruces, J. M. Barcelo-Ordinas, and M. Fiore, “Exact decoding probability under random linear network coding,” IEEE Communications Letters, vol. 15, no. 1, pp. 67–69, 2011. View at Publisher · View at Google Scholar · View at Scopus
  45. Z. Ma, F. C. A. Fernandes, and Y. Wang, “Analytical rate model for compressed video considering impacts of spatial, temporal and amplitude resolutions,” in Proceedings of the IEEE International Conference on Multimedia and Expo Workshops (ICMEW '13), pp. 1–6, July 2013. View at Publisher · View at Google Scholar · View at Scopus
  46. S. B. Kodeswaran and A. Joshi, “Content and context aware networking using semantic tagging,” in Proceedings of the 22nd International Conference on Data Engineering Workshops (ICDEW '06), p. 77, IEEE, Atlanta, Ga, USA, April 2006. View at Publisher · View at Google Scholar
  47. C. Tsilopoulos, G. Xylomenos, and G. C. Polyzos, “Are information-centric networks video-ready?” in Proceedings of the 20th International Packet Video Workshop (PV '13), pp. 1–8, San Jose, Calif, USA, December 2013. View at Publisher · View at Google Scholar · View at Scopus
  48. M. A. Pimentel-Niño, M. A. Vázquez-Castro, and I. Hernáez-Corres, “Perceptual semantics for video in situation awareness,” in Proceedings of the 9th International Conference on Systems and Networks Communications (ICSNC '14), pp. 11–16, October 2014.
  49. C. Yim and A. C. Bovik, “Evaluation of temporal variation of video quality in packet loss networks,” Signal Processing: Image Communication, vol. 26, no. 1, pp. 24–38, 2011. View at Publisher · View at Google Scholar · View at Scopus
  50. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004. View at Publisher · View at Google Scholar · View at Scopus