scispace - formally typeset
Search or ask a question

Showing papers on "Packet loss published in 2022"


Journal ArticleDOI
TL;DR: The FBQACC is simulated using the NS2 simulator and it gives an improvement in the average throughput FTLP and FEWPBRC by 7.1499% and 6.3327%, respectively and the proposed work gives a minimized average packet loss percentage compared to the existing congestion control methods.
Abstract: Wireless Multimedia Sensor Networks (WMSNs) are mostly affected by bottleneck issues, high packet loss, increased delay, and minimum throughput. One of the most effective schemes towards controlling the bottleneck on the web is traffic control. The WMSNs handle different types of data, hence QoS is essential to afford trustworthy as well as reasonable services towards these kinds of data. The existing congestion control methods, FTLP and FEWPBRC consider the frequency of packet transmission and decide on the output transmission rate of the sink. In the Fuzzy-Based QoS Alert Bottleneck Mechanism, the probability of congestion is predicted by using a fuzzy inference system with three special congestion indicators, and the traffic rate is adjusted based on the priority of the real-time and non-real-time applications. The FBQACC is simulated using the NS2 simulator and it gives an improvement in the average throughput FTLP and FEWPBRC by 7.1499% and 6.3327%, respectively. Similarly, FBQACC reduces average delay compared to FTLP and FEWPBRC by 11.074% and 7.8128%, respectively. The proposed work also gives a minimized average packet loss percentage compared to the existing congestion control methods.

18 citations


Journal ArticleDOI
TL;DR: In this paper, a minimum edge-shared vertex path between the source and the base station is created by updating neighbor information using Ant Colony Algorithm (ACA), and the body controller in the proposed system minimizes the data traffic entered into the network through the use of packet discard policy.

16 citations


Journal ArticleDOI
TL;DR: In this paper , a minimum edge-shared vertex path between the source and the BS is created by updating neighbor information using Ant Colony Algorithm (ACA), and the body controller in the proposed system minimizes the data traffic entered into the network through the use of packet discard policy.

16 citations


Journal ArticleDOI
TL;DR: In this article , a retransmission-assisted resource management (R-ARM) system is proposed for LoRaWAN mobile applications that experience frequent channel attenuation when the ED moves (e.g., pet tracking).
Abstract: LoRaWAN exhibits an essential feature, namely, the adaptive data rate (ADR), which has been recommended for the management of resources (e.g., the spreading factor and transmit power) of static end devices (EDs) based on channel conditions. Blind ADR (BADR) has been introduced for LoRaWAN mobile applications that experience frequent channel attenuation when the ED moves (e.g., pet-tracking). This channel condition leads to massive packet loss and retransmission, which significantly increases energy consumption. In this study, ADR and BADR are investigated in mobility environments, their limitations are highlighted, and a novel ADR “retransmission-assisted resource management (R-ARM)” system is proposed. The proposed R-ARM system operates concurrently on the ED and network server sides. This improves the network performance of the LoRaWAN. When compared to those of typical ADR approaches, the simulation results, in this case, show that R-ARM significantly enhances the packet success ratio and convergence period, and it lowers the energy consumption and packet loss ratio.

15 citations


Proceedings ArticleDOI
25 Oct 2022
TL;DR: In this paper , the authors investigate the user-perceived performance of Starlink and show that latency remains low and does not vary significantly under idle or lightly loaded links, while delay and packet loss increase slightly under load for both upload and download.
Abstract: With new Low Earth Orbit satellite constellations such as Starlink, satellite-based Internet access is becoming an alternative to traditional fixed and wireless technologies with comparable throughputs and latencies. In this paper, we investigate the user-perceived performance of Starlink. Our measurements show that latency remains low and does not vary significantly under idle or lightly loaded links. Compared to another commercial Internet access using a geostationary satellite, Starlink achieves higher TCP throughput and provides faster web browsing. To avoid interference from performance enhancing proxies commonly used in satellite networks, we also use QUIC to assess performance under load and packet loss. Our results indicate that delay and packet loss increase slightly under load for both upload and download.

10 citations


Journal ArticleDOI
TL;DR: In this paper , a cooperative routing protocol (CRP) was proposed to improve package delivery via end-to-end latency and increase the length of the network lifetime, which increased the rate of transmission of packets to the sink and mitigate packet loss.
Abstract: The wireless body area network (WBAN) has been proposed to offer a solution to the problem of population ageing, shortage in medical facilities and different chronic diseases. The development of this technology has been further fueled by the demand for real-time application for monitoring these cases in networks. The integrity of communication is constrained by the loss of packets during communication affecting the reliability of WBAN. Mitigating the loss of packets and ensuring the performance of the network is a challenging task that has sparked numerous studies over the years. The WBAN technology as a problem of reducing network lifetime; thus, in this paper, we utilize cooperative routing protocol (CRP) to improve package delivery via end-to-end latency and increase the length of the network lifetime. The end-to-end latency was used as a metric to determine the significance of CRP in WBAN routing protocols. The CRP increased the rate of transmission of packets to the sink and mitigate packet loss. The proposed solution has shown that the end-to-end delay in the WBAN is considerably reduced by applying the cooperative routing protocol. The CRP technique attained a delivery ratio of 0.8176 compared to 0.8118 when transmitting packets in WBAN.

9 citations


Journal ArticleDOI
TL;DR: In this article , a hybrid machine learning (ML)-based semi-empirical path loss (PL) model for LoRa wireless communication is proposed for monitoring water quality, combining existing wireless technologies with the aid of an air balloon to relay data over long distances in hilly terrain.
Abstract: This article proposes a novel Airborne Internet of Things Network (AIN) system architecture for monitoring water quality, combining existing wireless technologies with the aid of a low altitude platform (LAP) to relay data over long distances in hilly terrain. The proposed system consists of water quality sensors, smart utility network (SUN) devices, long range (LoRa) wireless devices, an LAP air balloon, and Wi-Fi devices. A measurement campaign was conducted to assess the proposed system, focusing on the communication link reliability and the LAP stability and robustness. Several constraints, such as payload limit and safe weather conditions, were also highlighted for operating the LAP with extensive and reliable coverage. On the other hand, characterizing the wireless channel has become a crucial parameter for planning and deploying Internet of Things (IoT) applications. Accordingly, this work proposes a novel hybrid machine learning (ML)-based semi-empirical path loss (PL) model for LoRa wireless communication. The results validated the proposed system’s effectiveness, unique characteristics, and capability to monitor water quality in a harsh environment. Results also revealed a significant difference in packet delivery rate (PDR) for different gateway height and spreading factor (SF) configurations. For instance, switching SF7 to SF12 increased PDR by 28.7%. Meanwhile, increasing gateway height increased PDR by 29.2% for similar SF configurations. The evaluation also revealed that none of the established PL models are suitable to represent harsh tropical environments. Finally, the proposed model achieved nearly 90% prediction accuracy for testing samples and 95% accuracy for training and overall measurement samples, vastly outperforming conventional models.

9 citations


Journal ArticleDOI
TL;DR: In this article, a disaggregated application-centric optical network (DACON) for data center infrastructures based on hybrid optical switches is proposed and experimentally demonstrated.
Abstract: To solve the issues of low resource utilization and performance bottleneck in current server-centric data center networks (DCNs), we propose and experimentally demonstrate a disaggregated application-centric optical network (DACON) for data center infrastructures based on hybrid optical switches. DACON achieves flexible provision and dynamic reconfiguration of hardware nodes exploiting the softwared-define networking (SDN) orchestration plane. A four-node SDN-enabled disaggregated prototype is implemented with a field-programmable-gate-array-based controller of hardware nodes and nanosecond optical switches, performing a minimal node-to-node network latency of 378.6 ns and zero packet loss. Based on the unmodified Linux kernel and two different applications (distributed computing and a Memcached database), the application runtime of the disaggregated prototype is investigated and compared with the server-centric architecture. Experimental results show that the disaggregated prototype performs better with Memcached database applications, achieving a 1.46× faster runtime than the server-centric network at a memory node access ratio of 0.9. Based on the customized control plane orchestrator and dynamic resource reallocation, the node-to-node latency is reduced by 21% when CPU nodes access memory nodes. The scalability of DACON is then numerically assessed based on experimentally measured parameters. Results show that the intra-rack node-to-node latency is less than 404.8 ns with a 6144-node network and memory node access ratio of 0.9. Finally, the cost and power consumption are also studied and compared with current DCN architectures. Results indicate DACON saves 13.4% of the cost of an interconnect network compared with current disaggregated architecture and consumes up to 31.1% less power with respect to server-centric DCN architectures.

8 citations


Journal ArticleDOI
TL;DR: A novel Distributed Tree-based Multicast Routing (DTMR) algorithm is proposed, which achieves network stability and reliability in real-time and outperforms the existing DTRBIP and EEMSFV protocols.

8 citations


Journal ArticleDOI
TL;DR: In this article , the Optimized Link State Routing Protocol (OLSR) is used to investigate the performance of VoIP applications in the VANET network, the network was tested before and after running the OLSR algorithm, the testing focused on Quality of Service (QoS) parameters like end-to-end delay, delay variation (jitter), and probability of packet loss between two moving hops through multi-hop Ad-hoc networks in different scenarios, using the ITU G.711 VoIP codec.
Abstract: Voice over Internet Protocol (VoIP) provides good services through Vehicular Ad-hoc Networks (VANET) platform. These networks encounter various challenges to support voice calls with acceptable Quality of Service (QoS). The Optimized Link State Routing Protocol (OLSR) is used to investigate the performance of VoIP applications in the VANET network. The network was tested before and after running the OLSR algorithm, the testing focused on Quality of Service (QoS) parameters like end-to-end delay, delay variation (jitter), and probability of packet loss between two moving hops through multi-hop Ad-hoc networks in different scenarios, using the ITU G.711 VoIP codec. After measured the end-to-end delay, jitter, and probability of packet loss for two nodes the algorithm decreased delay by 18.72%, while decreased jitter by about 20.42% and decreased packet loss by about 128.6%. However, The OLSR has shown initial good performance for four hops, and then added more hops the delay exceeded 400 ms which is not acceptable according to ITU-T recommendations. This was achieved by implementing a testbed to extract the desired results as opposed to using a simulation.

7 citations


Journal ArticleDOI
TL;DR: In this paper , the authors have evaluated the performance of LoRaWAN communication protocol for the implementation of the IoT (Internet of Things) nodes' communication in a forest scenario.
Abstract: The simultaneous advances in the Internet of Things (IoT), Artificial intelligence (AI) and Robotics is going to revolutionize our world in the near future. In recent years, LoRa (Long Range) wireless powered by LoRaWAN (LoRa Wide Area Network) protocol has attracted the attention of researchers for numerous applications in the IoT domain. LoRa is a low power, unlicensed Industrial, Scientific, and Medical (ISM) band-equipped wireless technology that utilizes a wide area network protocol, i.e., LoRaWAN, to incorporate itself into the network infrastructure. In this paper, we have evaluated the LoRaWAN communication protocol for the implementation of the IoT (Internet of Things) nodes’ communication in a forest scenario. The outdoor performance of LoRa wireless in LoRaWAN, i.e., the physical layer, has been evaluated in the forest area of Kashirampur Uttarakhand, India. Hence, the present paper aims towards analyzing the performance level of the LoRaWAN technology by observing the changes in Signal to Noise Ratio (SNR), Packet Reception Ratio (PRR) and Received Signal Strength Indicator (RSSI), with respect to the distance between IoT nodes. The article focuses on estimating network lifetime for a specific set of LoRa configuration parameters, hardware selection and power constraints. From the experimental results, it has been observed that transmissions can propagate to a distance of 300 m in the forest environment, while consuming approx. 63% less energy for spreading factor 7 at 2 dBm, without incurring significant packet loss with PRR greater than 80%.

Journal ArticleDOI
TL;DR: In this paper , an SDN enhanced P2P virtual reality (VR) live video streaming application NG-EPON multicast architecture based on 5G-enabled Tactile Internet (TI) and aims to improve the QoS, QoE and provide flexible services to the end-users.
Abstract: The IEEE 802.3ca Task Force group has recently approved the standardization of the next-generation Ethernet passive optical network (NG-EPON) to satisfy future bandwidth demands. NG-EPON architecture classifies the optical network unit (ONU) transmission to allow simultaneous Wavelength-Agile (WA-PON) transmission, which increases the network capacity (50 Gb/s) and flexibility. The recently launched 5G services and related functional standards, new capabilities, and quality of service (QoS) imposed on the optical transport network. The Peer-to-peer (P2P) based live streaming application is the most popular multimedia application in access networks, i.e., Facebook Live, YouTube Live, Twitch, and other streaming and video conferencing services are increasing, especially in global epidemics which requires a higher QoS. Hence, media-based applications require adequate infrastructure resources to meet customer demands such as high bandwidth, low latency, high quality audiovisual, live streaming, and services. Therefore, this paper proposes an SDN enhanced P2P virtual reality (VR) live video streaming application NG-EPON multicast architecture based on 5G-enabled Tactile Internet (TI) and aims to improve the QoS, QoE and provide flexible services to the end-users. Finally, we proposed an SDN-based TI-dynamic wavelength and bandwidth allocation (SD-TI-DWBA) mechanism. Simulation results show that our proposed architecture can improve the system performance and QoS metrics regarding packet delay, jitter, system throughput, and packet loss.


Journal ArticleDOI
01 Jan 2022
TL;DR: In this paper, the effects of traffic-induced delay and dropout on the finite-horizon quality of control of an individual stochastic linear time-invariant system, where quality-of-control is measured by an expected quadratic cost function, are analyzed.
Abstract: Transmission delay and packet dropout are inevitable network-induced phenomena that severely compromise the control performance of network control systems. The real-time network traffic is a major dynamic parameter that directly influences delay and reliability of transmission channels, and thus, acts as an unavoidable source of induced coupling among all network sharing systems. In this letter, we analyze the effects of traffic-induced delay and dropout on the finite-horizon quality-of-control of an individual stochastic linear time-invariant system, where quality-of-control is measured by an expected quadratic cost function. We model delay and dropout of the channel as generic stochastic processes that are correlated with the real-time network traffic induced by the rest of network users. This approach provides a pathway to determine the required networking capabilities to achieve a guaranteed quality-of-control for systems operating over a shared-traffic network. Numerical evaluations are performed using realistic stochastic models for delay and dropout. As a special case, we consider exponential distribution for delay with its rate parameter being traffic-correlated, and traffic-correlated Markov-based packet drop model.

Journal ArticleDOI
TL;DR: In this article , the authors evaluated the performance of three channel bonding techniques from the IEEE 802.11 family of standards considering frame transmission delay and packet loss ratio, and provided a recommendation for selecting both the most suitable channel access method and primary channel.
Abstract: In recent years, Vehicle-To-Everything (V2X) applications have been actively developing, and the Quality of Service (QoS) requirements are becoming more stringent. To support V2X applications, IEEE developed the IEEE 802.11p standard in 2010. However, systems based on this standard fail to fulfill requirements on very low frame transmission delay and packet loss ratio imposed by modern V2X applications, such as applications for autonomous driving and platooning. To satisfy new requirements, IEEE has launched a new IEEE 802.11bd project to design the next generation of IEEE 802.11p. An important feature of IEEE 802.11bd is the channel bonding technique, which allows transmitting data in two adjacent channels simultaneously. Thus, it increases data rates and may reduce delays and packet loss ratio. The current version of IEEE 802.11bd specifies two channel bonding techniques, which differ from that used in modern Wi-Fi networks. This work evaluates the performance of the three aforementioned techniques from the IEEE 802.11 family of standards considering frame transmission delays and packet loss ratio. With rigorous simulations, it is shown that, in most cases, the channel bonding techniques highly decrease the percentage of both IEEE 802.11bd and legacy stations with unsatisfied QoS requirements on delays and packet loss ratio. Unfortunately, sometimes, the IEEE 802.11bd channel bonding techniques significantly worsen performance. Moreover, the work highlights that the choice of the primary channel for channel bonding techniques significantly affects the network performance. As a result, the paper provides a recommendation for selecting both the most suitable channel access method and primary channel.

Proceedings ArticleDOI
Nan Li, Xiguang Zheng, Chen Zhang, Liang Guo, Bin Yu 
18 Sep 2022
TL;DR: This work presents a low-delay multi-loss based neural PLC system consisted by a signal loss, a perceptual loss and an ASR loss ensuring good speech quality and automatic speech recognition compatibility.
Abstract: Real-time teleconferencing has become one of the essential parts in our daily life. While packet loss during real-time data transmission is unavoidable, traditional signal processing based Packet Loss Concealment (PLC) techniques have been devel-oped in recent decades. In recent years, deep learning based approaches have also proposed and achieved state-of-the-art PLC performance. This work presents a low-delay multi-loss based neural PLC system. The multi-loss is consisted by a signal loss, a perceptual loss and an ASR loss ensuring good speech quality and automatic speech recognition compatibility. The proposed system was ranked 1 st place in INTERSPEECH 2022’s Audio Deep Packet Loss Concealment Challenge.

Journal ArticleDOI
TL;DR: In this paper, the authors explored different applications of patient-centric IoHT architectures to draw a realistic resource-limited topological layout of IoHT for congestion estimation, and proposed an effective buffer-loss estimation model based on the Queuing Theory to determine the number of packets lost at the node's buffer.

Journal ArticleDOI
TL;DR: In this paper , the authors present an overview of how Veins VANET simulator simulates WiFi packet loss and explain the physical parameters set by the user in Veins and how VeINS then uses these parameters to simulate the effects of attenuation and packet loss.

Journal ArticleDOI
01 Jan 2022-Sensors
TL;DR: This paper proposes a Machine Learning method to predict whether PD is required at a specific data transmission to successfully send a URLLC message, and shows that the prediction model reduced the number of packets sent with PD by 81% while maintaining the same level of latency as a static PD technique.
Abstract: The fifth-generation (5G) network is presented as one of the main options for Industry 4.0 connectivity. To comply with critical messages, 5G offers the Ultra-Reliable and Low latency Communications (URLLC) service category with a millisecond end-to-end delay and reduced probability of failure. There are several approaches to achieve these requirements; however, these come at a cost in terms of redundancy, particularly the solutions based on multi-connectivity, such as Packet Duplication (PD). Specifically, this paper proposes a Machine Learning (ML) method to predict whether PD is required at a specific data transmission to successfully send a URLLC message. This paper is focused on reducing the resource usage with respect to pure static PD. The concept was evaluated on a 5G simulator, comparing between single connection, static PD and PD with the proposed prediction model. The evaluation results show that the prediction model reduced the number of packets sent with PD by 81% while maintaining the same level of latency as a static PD technique, which derives from a more efficient usage of the network resources.

Journal ArticleDOI
TL;DR: The simulation results validate that the proposed upper confidence bound (UCB)-based dynamic CoAP mode selection algorithm can flexibly balance the tradeoff between packet-loss ratio and transmission delay as well as satisfy the differentiated QoS in distribution IoT.
Abstract: Lightweight constrained application protocol (CoAP) has emerged as a common communication protocol for resource-constrained equipment in distribution internet of things (IoT). CoAP introduces two modes for data transmission, i.e., non-confirmed mode for reducing transmission delay and confirmed mode for reducing packet-loss ratio, which can be dynamically selected to satisfy the service requirements. However, there are still some challenges in dynamic CoAP mode selection, including incomplete information and differentiated quality of service (QoS) requirements of distributed IoT services. In this paper, we propose a upper confidence bound (UCB)-based dynamic CoAP mode selection algorithm for data transmission to address these challenges. The simulation results validate that, compared with the fixed mode selection algorithm, the proposed algorithm can flexibly balance the tradeoff between packet-loss ratio and transmission delay as well as satisfy the differentiated QoS in distribution IoT.

Journal ArticleDOI
TL;DR: In this article , the authors explored different applications of patient-centric IoHT architectures to draw a realistic resource-limited topological layout of IoHT for congestion estimation, and proposed an effective buffer-loss estimation model based on the Queuing Theory to determine the number of packets lost at the node's buffer.

Journal ArticleDOI
TL;DR: In this paper , the quality of service (QoS) performance of wireless networks that integrate light fidelity (LiFi) and wireless fidelity (WiFi) was studied for hybrid LiFi and WiFi networks.
Abstract: This work studies the quality of service (QoS) performance of wireless networks that integrate light fidelity (LiFi) and wireless fidelity (WiFi). While the hybrid network is potential for improving network capacity, load balancing becomes essential and challenging due to the nature of heterogeneous access points (APs). A number of studies have been conducted to address this issue, focusing on maximising the network capacity with user fairness constraints. However, in practice, QoS metrics including packet loss ratio and latency are important to network services. In this paper, QoS-driven load balancing is studied for hybrid LiFi and WiFi networks (HLWNets) in two scenarios: single-AP association (SA) and multi-AP association (MA). In each case, an optimisation problem is formulated to minimise the packet loss ratio and latency, and a low-complexity iterative algorithm is proposed to solve the problem. Results show that the novel methods, especially MA, can effectively balance the traffic loads among the APs and improve the QoS performance. In addition, the more subflows the better performance MA provides. Targeting the same level of QoS, MA can achieve a system throughput up to 160% higher than the signal strength strategy and 130% higher than the proportional fairness load balancing.

Journal ArticleDOI
TL;DR: In this paper , a game-theoretic auction-based framework is proposed to find the priority of relays for a given D2D pair based on the LOS blockage probabilities.


Journal ArticleDOI
TL;DR: In this article , the authors proposed an upper confidence bound (UCB)-based dynamic CoAP mode selection algorithm for data transmission to address incomplete information and differentiated quality of service (QoS) requirements of distributed IoT services.
Abstract: Lightweight constrained application protocol (CoAP) has emerged as a common communication protocol for resource-constrained equipment in distribution internet of things (IoT). CoAP introduces two modes for data transmission, i.e., non-confirmed mode for reducing transmission delay and confirmed mode for reducing packet-loss ratio, which can be dynamically selected to satisfy the service requirements. However, there are still some challenges in dynamic CoAP mode selection, including incomplete information and differentiated quality of service (QoS) requirements of distributed IoT services. In this paper, we propose a upper confidence bound (UCB)-based dynamic CoAP mode selection algorithm for data transmission to address these challenges. The simulation results validate that, compared with the fixed mode selection algorithm, the proposed algorithm can flexibly balance the tradeoff between packet-loss ratio and transmission delay as well as satisfy the differentiated QoS in distribution IoT.

Journal ArticleDOI
01 Jun 2022-Sensors
TL;DR: A novel Deep Learning (DL)-based Cooperative communication channel model for Wireless Underground Sensor Networks for accurate and reliable monitoring in hostile underground locations and demonstrates some superiority over other methods in terms of its better energy utilization.
Abstract: Wireless Underground Sensor Networks (WUSNs) have been showing prospective supervising application domains in the underground region of the earth through sensing, computation, and communication. This paper presents a novel Deep Learning (DL)-based Cooperative communication channel model for Wireless Underground Sensor Networks for accurate and reliable monitoring in hostile underground locations. Furthermore, the proposed communication model aims at the effective utilization of cluster-based Cooperative models through the relay nodes. However, by keeping the cost effectiveness, reliability, and user-friendliness of wireless underground sensor networks through inter-cluster Cooperative transmission between two cluster heads, the determination of the overall energy performance is also measured. The energy co-operative channel allocation routing (ECCAR), Energy Hierarchical Optimistic Routing (EHOR), Non-Cooperative, and Dynamic Energy Routing (DER) methods were used to figure out how well the proposed WUSN works. The Quality of Service (QoS) parameters such as transmission time, throughput, packet loss, and efficiency were used in order to evaluate the performance of the proposed WUSNs. From the simulation results, it is apparently seen that the proposed system demonstrates some superiority over other methods in terms of its better energy utilization of 89.71%, Packet Delivery ratio of 78.2%, Average Packet Delay of 82.3%, Average Network overhead of 77.4%, data packet throughput of 83.5% and an average system packet loss of 91%.

Journal ArticleDOI
TL;DR: The enhanced BBR congestion control (eBCC) algorithm is proposed, which improves the BBR algorithm in two aspects: 10.87% higher throughput and 74.58% lower packet loss rate in the low-bandwidth scenario and 8.39% higher fairness in the multi-stream scenario.
Abstract: Congestion control is a fundamental technology to balance the traffic load and the network. The Internet Engineering Task Force (IETF) Quick UDP Internet Connection (QUIC) protocol has flexible congestion control and at the same time possesses the advantages of high efficiency, low latency, and easy deployment at the application layer. Bottleneck bandwidth and round-trip propagation time (BBR) is an optional congestion control algorithm adopted by QUIC. BBR can significantly increase throughput and reduce latency, in particular over long-haul paths. However, BBR results in high packet loss in low bandwidth and low fairness in multi-stream scenarios. In this article, we propose the enhanced BBR congestion control (eBCC) algorithm, which improves the BBR algorithm in two aspects: (1) 10.87% higher throughput and 74.58% lower packet loss rate in the low-bandwidth scenario and (2) 8.39% higher fairness in the multi-stream scenario. This improvement makes eBCC very suitable for IoT communications to provide better QoS services.

Book ChapterDOI
01 Jan 2022
TL;DR: The proposed trust community and optimized RPL is best for the trust-based data transmission in social Internet of things network by having high alive nodes of 90% and time delay of 30 s for hop localization.
Abstract: Internet of things plays a major role in all the applications that are utilized for automating the environment. In this, the searching of devices in the larger network is a tedious process. Further, it became easier by defining the social relationship among the devices in the network. Social Internet of things is the combination of real-world network and multiple devices in the Internet of things. In social Internet of things, the user can access information globally at any place based on its ID. But, the accessing of this ID in the large database is difficult. Several works were carried out to locate the ID of the devices in the social Internet of things [SIoT]. Community and effective search-based algorithm is one of the methods for effective localization of devices in the social Internet of things as the conventional method for device localization. The Routing protocol for Low power and Lossy networks (RPL) able to increase lifetime of network with high packet loss. In existing, the localization and routing is performed separately. This problem is overcome with the proposed trust-based community nodes and optimized RPL in social Internet of things network. Here, the community is formed with the combination of high trust and low trust nodes of similar taste. These trusted node-based community formation helps to locate the trusty nodes in the transmission process. The dragon fly optimization is used to improve the RPL performance by selecting trusty nodes as next hop in the routing process. Due to the combination of trust community and optimized RPL, the lifetime and success rate of the data transmission is increased by 90% as compared to the existing preference-based device localization and traditional RPL protocol. The whole process is realized with the help of MATLAB R2018a in windows 10 environment. Therefore, the proposed trust community and optimized RPL is best for the trust-based data transmission in social Internet of things network by having high alive nodes of 90% and time delay of 30 s for hop localization.

Proceedings ArticleDOI
10 Apr 2022
TL;DR: In this paper , an opportunistic random redundant retransmission (OR3) algorithm is proposed to mitigate the mismatch problem between loss recovery and high-speed TCP advancements such as non-loss-based congestion control.
Abstract: The emerging applications including cloud AR/VR gaming, ultra-high definition (UHD) streaming, Metaverse, etc. imply the demand for ultra-high bandwidth transmission in the wide area network (WAN). During the past four decades, a great number of high-speed TCP variants have been proposed to improved the throughput for WAN transmission. This paper conducts a measurement study on loss recovery for high-speed transmission, and exposes that receive buffer starvation is a result of capability mismatch between loss recovery and high-speed TCP advancements such as non-loss-based congestion control. To mitigate this mismatch problem, an opportunistic random redundant retransmission (OR3) algorithm, as well as its TCP implementation TCP-OR3, is proposed. OR3 accelerates loss recovery by minimizing the maximum retransmission times of each packet. Experiment results shows that TCP-OR3 alleviates receive buffer starvation and tackles the bottleneck of the legacy approaches such as parallel TCP in the case of high-speed transmission.

Journal ArticleDOI
TL;DR: In this article , three different transmission strategies, namely caching, network coding, and repetition enabled transmission in the user plane (UP) of mobile backhaul for network operators to prevent such performance degradation, are proposed and compared.
Abstract: Avoiding problems such as packet loss in the transport network is crucial for mobile network providers to offer high-quality services reliably and without interruptions. In this paper, we propose and compare three different transmission strategies, namely Caching, Network Coding (NC) and Repetition enabled transmission in the User Plane (UP) of mobile backhaul for network operators to prevent such performance degradation. In the proposed NC-enabled transmission method, NC provides robustness to transport network failures such that no further retransmission is required by the User Equipment (UE) compared to conventional approaches where UE applications perform retransmissions. The proposed scheme requires only a minor modification to the packet structure of the UP protocol, which requires a small development effort and no new extensions to the current UE standard features. We also discuss their placement in the O-RAN protocol stack and in the core network, and propose a new architecture that can utilize caching, repetition and NC features in the mobile network architecture. Our simulation results show that an exact 1% packet loss ratio in the backhaul link results in an additional total transmission time of 59.44% compared to the normal GPRS Tunneling Protocol-User Plane (GTP-U) transmission. Applying NC at a rate of 1% and 2% reduces this value to 52.99% and 56.26%, respectively, which is also better than the total transmission time of some previously studied dynamic replication schemes while keeping the bandwidth utilization at low rates. On the cache side, a reduction in latency of about 20% can be achieved with a cache size of 100 MB. At the end of the paper, we summarize some of the benefits and limitations of using these three strategies in UP of mobile backhaul networks.